Data Science Training Ahmedabad
08 November 2023

Want to Improve your Data Science skills? Join a Certification program for Career Growth

If you are interested in making a career in data science, then there is a high possibility to get a lucrative job role. There is a growing demand for skilled professionals in the IT industry. More and more business corporates are in need of data scientists who have updated competence in analyzing data. It can be standard or complex sets that help companies make more informed decisions by interpreting the data. Therefore, you should have Data Science Certification in Ahmedabad that proves your knowledge in data science knowledge and in which cases you can use them.

How joining Data Science Certification programs can be valuable?

Having a data science certificate might not be a comprehensive solution for all your IT skills needs. But if you have a career in the IT sector, it can help you get a better job role while growing your skills. You can grow your experience in understanding various data sets and stay updated on how to use and comprehend data that are required for current IT trends. So, you should improve your qualifications as per the needs of your job role as a data analyst. Data scientists can also gain more skills to use the right algorithm for business data training with Microsoft Azure Training & Certification in Ahmedabad. If you work with data on the AWS platform, you can get benefits from the AWS Security Training Course in Ahmedabad. You can learn about working with a range of cloud-based services.

Today, businesses in various industries depend more on data scientists to handle the increasing volume of data created and compiled. This is where data science can play an increasingly important role in a variety of sectors, offering a plethora of career possibilities. All you have to do is have the proper qualifications to become a data scientist. Hence, taking data science certification programs is worthwhile. HighSkyIT Solution offers data science certification sources that enable you to possess skills on both Azure and AWS. So you can effectively tackle all your data science projects.

Conclusion:

Data science is a dynamic field in the IT realm that offers chances for study and job advancements with ever-evolving technology and approaches. Hence, you should enroll in a data science certification program that enables you to stay current on the newest tools and trendy strategies. It will help you become a competitive and productive professional in the industry. Whether you want to become a data analyst or a big data engineer, to meet the need for evolving skills, taking on these certification courses can be helpful in career growth.

10 October 2023

What is Nodejs And install Nodejs

What is Node.js

An open-source server-side runtime environment for JavaScript called Node.js enables you to execute JavaScript code on the server. It is made for creating scalable and effective network applications, and it is widely used to create web applications, APIs, and other server-side programs.

1 JavaScript Runtime:- Node.js enables you to execute JavaScript code on the server, whereas JavaScript is typically associated with web browsers for client-side scripting.

2 Single-Threaded:- Although Node.js applications are single-threaded, they may effectively manage several concurrent tasks by using callbacks and asynchronous activities. For some applications, this may lead to better performance.

3 Event-Driven and Non-Blocking:- An event-driven, non-blocking I/O mechanism is the foundation of Node.js. This indicates that it can manage numerous connections simultaneously without delaying the execution of any code. Applications that demand a lot of concurrency and real-time communication are especially well suited for it.

4 Package Manager (npm):- The npm (Node Package Manager) ecosystem of open-source libraries and modules that come with Node.js allows developers to expand the capabilities of the framework and make the process of creating applications more straightforward. ion.

5 V8 JavaScript Engine:- The V8 JavaScript engine, created by Google and renowned for its great performance, is used by Node.js. V8 is quick and effective because JavaScript code is compiled into machine code. n.

6 Cross-Platform:- Node.js is extremely portable since it supports many different operating systems, including Windows, macOS, and many Unix-like platforms.

7 Large and Active Community:- A large number of resources, libraries, and tools are available for Node.js development because of the dynamic and active developer community that supports it.

The development of web servers and web applications, real-time applications like chat programs and online games, the creation of APIs, and the creation of command-line tools are just a few examples of common Node.js use cases. Its popularity has increased in part as a result of its adaptability in creating various applications and its ability to handle asynchronous I/O activities effectively.

install Node js

The Node.js official website provides a yum repository that must first be enabled on your system. Additionally, you require development tools to create native add-ons that may be installed on your system.

The default repositories of Rhel include a version of Node.js that can be used to deliver a consistent user experience across several platforms. Version 12.22.9 is what is currently available in the repository. While not the most recent version, this one should be reliable and sufficient for fast language testing.

1 You can use the yum package manager to obtain this version. First, update your local package index by typing:

yum update
yum install nodejs

To confirm installation, click Y when prompted. If you are asked to restart any services, hit ENTER to proceed with the default settings. Ask the node for its version number to confirm the installation was successful:

node -v

This is all there is to getting started with Node.js if the package in the repositories meets your needs. You should typically install npm, the Node.js package manager, as well. You can accomplish this by using yum to install the npm package:

2 You don’t need to install npm individually because the NodeSource nodejs package includes both the node binaries and npm.

By using apt and the NodeSource PPA, you have now successfully installed Node.js and npm. The installation and management of several Node.js versions are covered in the section that follows.

yum install npm

 

3 The Node Version Manager is Used to Install Node

The Node Version Manager, or nvm, is a further flexible method of installing Node.js. You can simultaneously install and maintain numerous independent Node.js versions and their corresponding Node packages with this piece of software.

The project’s GitHub page can be accessed to learn how to install NVM on a Rhel 9 computer. The README file is displayed on the main page. Copy the curl command from there. This will provide you with the installation script’s most recent version.

It is usually a good idea to audit the script to ensure it isn’t doing anything you disagree with before papering the command through to bash. Remove the | bash element from the end of the curl command to accomplish that:

curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.39.1/install.sh

 

Make sure you understand the adjustments it is making by taking a look. Run the command once more with | bash appended after you are finished. As of right now, the script can be downloaded and run by typing: The URL you use will change based on the most recent version of nvm.

curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.39.1/install.sh | bash

The nvm script will be installed on your user account as a result. You need to source your.bashrc file first before using it:

source ~/.bashrc

You can now inquire NVM about the Node versions that are offered:

nvm list-remote

The list is pretty lengthy! Any of the release versions you see can be typed to install a particular version of Node. For example, you can enter the following to obtain version v16.14.0 (another LTS release):

nvm install v16.14.0

If you type: you can view the various versions you have installed.

nvm list

 

28 September 2023

Kubernetes Cluster Installation on RHEL 9

Kubernetes is an open-source container orchestration platform created to automate the deployment, scaling, administration, and orchestration of containerized applications. It is frequently shortened as K8s (K-8 characters between “K” and “s”). The Cloud Native Computing Foundation (CNCF) now maintains it after Google initially built it. Kubernetes is an effective platform for scalable, high-performance management of containerized applications. Here are some essential Kubernetes ideas and elements:

1 Container Orchestration: A platform for automating the deployment and maintenance of containerized applications is offered by Kubernetes. Containers are application-running environments that are compact, portable, and reliable. Based on resource usage and application needs, Kubernetes aids in ensuring that containers are deployed and scaled appropriately.

2 Kubectl: The command-line tool used to communicate with a Kubernetes cluster is called Kubectl. It enables users to build, examine, and control clusters and resources for Kubernetes.

3 Cluster Management: With a master node and numerous worker nodes, Kubernetes functions as a cluster. The cluster is managed and controlled by the master node, and containerized applications are run on the worker nodes. Due to its distributed architecture, high availability, and fault tolerance are guaranteed.

4 Containers: Applications are packaged and operated by Kubernetes in isolated, repeatable environments using container runtimes like Docker. Containers offer consistency between many settings, from production to development.

5 Pods: The Kubernetes term for the smallest deployable unit is “pod.” One or more containers in the same network and storage namespace can make up a pod. Co-located and co-scheduled on the same host, containers within a pod can easily communicate with one another.

6 Services: With Kubernetes, load balancing and the network are abstracted for applications utilizing services. Services give users a consistent virtual IP address and DNS name that may be used to direct traffic to a collection of pods. Because of this, apps can scale horizontally while preserving a constant network endpoint.

7 Replication Controllers and Replica Sets: These controllers guarantee that an agreed-upon number of pod replicas are active at all times. According to required replica counts, they are in charge of scaling up or down pods.

RHEL 9 had not yet been released as of my most recent knowledge update in September 2021, and depending on the version of RHEL you are running, the procedure for installing Kubernetes may differ. The most recent installation instructions are in the official RHEL 9 and Kubernetes documentation, which I strongly advise consulting. However, I can provide you with a rough breakdown of the procedures needed to install Kubernetes on RHEL:

Installing Kubernetes on RHEL Step By Step:

1 Docker (Container Runtime) installation:

Click ON The link to install the Docker:-

What Is Docker? How To Install RHEL 9

2 Disable the firewall and SELinux (optional):-
Since firewalls and SELinux might cause problems for Kubernetes, it’s frequently advised to turn them off. However, you need to set up SELinux and firewall rules for Kubernetes properly in a production environment. To momentarily turn off the firewall and SELinux:

( 1 ) Open the SELinux configuration file: /etc/selinux/config

[root@server ~]# vim /etc/selinux/config

( 2 ) Locate the following line:-

SELINUX=enforcing

( 3 ) Disabled should now be the value.

SELINUX=disabled

Close the file after saving your modifications.

( 4 ) SELinux is indefinitely disabled upon the subsequent reboot. Execute the following command to dynamically disable it prior to rebooting:

[root@server ~]# setenforce 0

3 Next Firewalld Disable and Stop 

[root@server ~]# systemctl stop firewalld.service
[root@server ~]# systemctl disable firewalld.service

3 Create a New Repository Kubernetes:- 

Installing Kubernetes components may be done using the official RHEL Kubernetes repository. Install Kubernetes after adding the repository:

tee /etc/yum.repos.d/kubernetes.repo <<EOF
[kubernetes]
name=Kubernetes
baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-\$basearch
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg
https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg
EOF

4 The Kubernetes components are now ready for installation:-

[root@server ~]# yum install -y kubeadm kubelet kubectl

5 Start the Kubernetes services by enabling them:

[root@server ~]# systemctl start kubelet
[root@server ~]# systemctl enable kubelet

 

27 September 2023

How To Create Auto Scaling In [ AWS ]

It takes a few steps to set up auto-scaling in AWS, and it’s commonly used to dynamically change the number of Amazon EC2 instances in a group to match shifting workloads. Here is a step-by-step tutorial for setting up auto-scaling on AWS:

1 Logging into the AWS Console:
Using the login information for your AWS account, access the AWS Management Console.

How TO  CERATE (AMI)
How to take AMI of EC2 and launch new EC2 using AMI

2 Selecting or Building an Amazon Machine Image (AMI)
The configuration of the EC2 instances you want to launch might be represented by an existing AMI or by a custom one that you generate.

3 Create a Launch Configuration:
1 Go to the EC2 Dashboard.
2 Select “Launch Configurations” from the left navigation pane’s “Auto Scaling” section.
3 The “Create launch configuration” button should be clicked.


Launch template name
= highsky_template
Template version description = template_highsky

4 Choose the AMI that you want.

Click = My AMIs
And Click Amazon Machine Image (AMI) [ Image Name ] = auto_image

5 Set up the instance type, key pair, security groups, and, if necessary, any user data scripts.

Choose the instance type = t2.micro

Choose you’re  Key =   – – – – – – 

Choose you’re Network Settings 

6 After reviewing the settings, click “Create launch configuration.

                             Create launch template

4 Create an Auto Scaling Group:
1 Go to the EC2 Dashboard and select “Auto Scaling Groups” from the left navigation pane after creating the launch configuration.


2 Press the “Create an auto-scaling group” button.

Auto Scaling group name = Auto_scaling_group
Launch template = highsky_template
Click = Next

3 The launch configuration you made in the previous stage should be selected.

VPC = ( Default VPC )
Availability Zones and Subnets = [ yure Choose )w
And Click = Next 

4 Configure advanced options – optional: [ Choose a load balancer to distribute incoming traffic for your application across instances to make it more reliable and easily scalable. You can also set options that give you more control over health check replacements and monitoring.]

I’m Choose = No load balancer

5 Health checks [ Health checks increase availability by replacing unhealthy instances. When you use multiple health checks, all are evaluated, and if at least one fails, instance replacement occurs.]

  Health check grace period = 180 Minute
  And Click = Next

6 Set the group’s desired capacity, minimum, and maximum instance counts.

Desired capacity = 1
Minimum capacity = 1
Maximum capacity = 2
And Click = Next

6 Set Up Notifications (Optional):
Notifications can be set up to notify you of scaling events. Email, SMS, and other destinations can receive these updates via Amazon SNS (Simple Notification Service).

Click  =  Next 

7 Test Auto Scaling:
1 Manually start scaling events by simulating traffic or load spikes to make sure your system behaves as you anticipate.
2 Watch how the Auto Scaling group changes the number of instances it has based on the policies you’ve set..

Click = Next

8 Monitoring and upkeep: 
1 Keep a close eye on the performance of your Auto Scaling group and modify scaling rules as necessary to meet your application’s needs.
2 Your instances’ health should be monitored, and any sick instances should be replaced immediately..

And Click = Create Auto Scaling groups 

Check-in Instances

                            Successfully Create Auto Scaling 

 

Docker Certification in Ahmedabad
15 September 2023

Get Your Docker Certification Demystified For Container Mastery

Do you want to get your Docker certification to get an industry-recognised credential? To get recognition, you must pass the Docker Certified Associate (DCA) exam. It’s time to start with a specific course to improve your docker skills. Courses for Docker Certification in Ahmedabad are available at competitive prices; along with professionals guide the candidates. Let’s know more about the certification course.

What You Will Achieve with the Certification Course?

  • Digital certificate and Docker Certified Associate logo.
  • Recognition of Docker skills with official Docker credentials.
  • Accessthe Docker Certified professional network.

While preparing for your Docker certification exam, you have to cover major concepts related to Docker skills to become a proficient developer, application architect,and system administrator. Here are the concepts you will cover;

Running Containerised Applications

You will learn to run containerised appsfrom pre-existing images. This concept will help you to improve your programming and development skills by enabling you to spin up dev environments. There are centres for DevOps Online Training Ahmedabad where you can learn this concept.

Deploying Images in the Cluster

Another major concept where you can learnto achieve continuous delivery is by deploying images in the cluster in the form of containers.

Installation and Maintenance of Docker platform

This concept will provide you with a clear insight into the Docker platform. Here, you will learn to install and operate the platform. Moreover, you will also get an idea of its maintenance and upgrades. It will provide you with an insight into the internals of Docker.

Configuration and Troubleshooting

In this concept, you will learn to configure and troubleshoot the Docker engine. There are prominent Cloud Computing Certifications Ahmedabad that also offer Docker certification courses, where all these concepts are covered. When you dive deep into the core topics of configuration and troubleshooting, you will cover topics such as Orchestration, Installation and Configuration, Storage and Volumes, Image Creation, Management, and RegistrySecurity and Networking.

Other Concepts of Container Mastery

There are also other concepts to cover in your docker platform, such as understanding triage issue reports from the stakeholders and resolving them. Knowledge of new Docker environments and performing general maintenance. Also, you will learn to migrate traditional applications to containers. This concept will help to migrateyour existing apps as Docker containerised apps. You can consult Ansible Training Ahmedabad to learn about the Docker certification.

These are the major concepts covered in Docker certification courses. To know more about the course, DCA exam, and concepts, get in touch with HighSkyIT Solution.

10 August 2023

how to create S3 bucket using Terraform

To use Terraform to construct an Amazon S3 bucket, you must define an appropriate resource block in your Terraform setup. Here’s a step-by-step tutorial on creating an S3 bucket with Terraform:

1 Configure AWS Credentials:
Before you continue, make sure you have your AWS credentials set up. You can use the AWS CLI aws configure command or specify them as environment variables.

2 Follow these steps to create a Terraform configuration:
Create a.tf file (for example, main.tf) to define your Terraform setup.

3 Define the S3 Bucket:
Add the following Terraform code to your main.tf file to define an S3 bucket resource:

terraform {
  required_providers {
    aws = {
        source = "hashicorp/aws"
        version = "5.8.0"
    }
  }
}

provider "aws" {
  region     = "ap-south-1"
  access_key = "Your_Access_Key"
  secret_key = "Your_Secrt_Key"

}

resource "aws_s3_bucket" "bucket" {
  bucket = "highskybucket"

  tags = {  
    Name        = "My bucket"
  }
}

Replace “highskybucket” with your S3 bucket’s unique name. Bucket names must be globally distinct throughout AWS.

4 Launch Terraform:
To launch Terraform, browse to the directory containing your Terraform configuration file and execute the following command:

 terraform init

5 Plan the Configuration:
In HashiCorp Terraform, the terraform plan command generates an execution plan outlining the modifications Terraform will make to your infrastructure based on your existing configuration. Without actually making the changes, it demonstrates to you what steps Terraform will take to create, update, or remove resources, for example. By doing so, you may examine and confirm the modifications before implementing them in your infrastructure.

terraform plan

6 Apply the Configuration:
Run the following command to create the AWS s3 Bucket:

terraform apply

7 Review and Confirm:
Terraform will display a plan of what it aims to build. After reviewing the plan, type yes to confirm and create the AWS s3 Bucket.

Output On AWS Infra

And Go To s3 Service 

 

How To Create IAM User and assign policy to user by Terraform

How to create ec2 instance using terraform

09 August 2023

How To Create IAM User and assign policy to user by Terraform

 To use Terraform to establish an AWS user, use the aws_iam_user resource given by the AWS provider. Here’s a step-by-step tutorial for creating an AWS user with Terraform.

1 Configure AWS Credentials:
Make sure you have your AWS credentials set up before you begin. You may either specify them as environment variables or use the AWS CLI aws configure command.

2 Create a Terraform configuration by following these steps:
To define your Terraform setup, create a.tf file (for example, main.tf).

3 Create an AWS User Resource:
To define the AWS user resource, add the following code to your main.tf file:

terraform {
  required_providers {
    aws = {
      source  = "hashicorp/aws"
      version = "~> 5.0"     }
  }
}

provider "aws" {
  region     = "ap-south-1"
  access_key = "your_Access_Key"
  secret_key = "Your_Secret_Key"

}


resource "aws_iam_user" "example_user" {
  name = "nitin_user"
}


resource "aws_iam_access_key" "kye1" {
  user = aws_iam_user.example_user.id

}


output "secret_key" {
  value     = aws_iam_access_key.kye1.secret
  sensitive = true
}


output "access_key" {
  value = aws_iam_access_key.kye1.id

}


resource "aws_iam_policy_attachment" "test-attach" {
  name       = "test-attachment"
  users      = [aws_iam_user.example_user.name]
#   roles      = [aws_iam_role.role.name]
#   groups     = [aws_iam_group.group.name]
  policy_arn = "arn:aws:iam::aws:policy/ReadOnlyAccess"
}

4 Initialize Terraform:
To start Terraform, navigate to the directory containing your Terraform configuration file and run the following command:

 terraform init

5 Plan the Configuration:
In HashiCorp Terraform, the terraform plan command generates an execution plan outlining the modifications Terraform will make to your infrastructure based on your existing configuration. Without actually making the changes, it demonstrates to you what steps Terraform will take to create, update, or remove resources, for example. By doing so, you may examine and confirm the modifications before implementing them in your infrastructure.

terraform plan

6 Apply the Configuration:
Run the following command to create the AWS user:

terraform apply

7 Review and Confirm:
Terraform will display a plan of what it aims to build. After reviewing the plan, type yes to confirm and create the AWS user.

Output On AWS Infra

And Check Policy 

How to create ec2 instance using terraform

Linux Certification Ahmedabad
19 July 2023

Continuing Education with Red Hat Staying Ahead in Open Source Technologies

In today’s rapidly evolving digital landscape, staying ahead in open-source technologies is essential for professionals seeking to excel in the field. With the vast popularity and significance of Linux administration and Red Hat technologies, it becomes crucial to equip oneself with the necessary skills and knowledge. If you are based in Ahmedabad, India, you’re in luck. A leading training provider offers top-notch Red Hat Training Course & Certification Ahmedabad designed to enhance your proficiency and open doors to exciting career opportunities.

Some features of Linux Administration with Online Classes in Ahmedabad

  • Comprehensive Curriculum:

This course provides a comprehensive curriculum that covers all aspects of managing and maintaining Linux-based systems. From basic concepts to advanced topics, you’ll gain a deep understanding of Linux architecture, command-line operations, user management, file systems, networking, security, and more. The curriculum is designed to equip you with the skills to handle real-world scenarios in Linux environments.

  • Flexibility and Convenience:

One of the primary advantages of online classes is the flexibility they offer. Whether you’re a working professional or a student, you can access the course materials and lectures at a time that suits you best. Companies like Highsky IT Solutions allow you to balance your learning with other commitments, making it convenient for individuals with busy schedules.

  • Interactive Learning Experience:

Engaging and interactive learning experiences are essential for effective comprehension and skill development. Through virtual labs, practical exercises, quizzes, and discussion forums, you’ll have hands-on opportunities to apply your knowledge, collaborate with peers, and seek guidance from experienced instructors.

  • Experienced Instructors:

To ensure a high-quality learning experience, Linux Administration Online Classes Ahmedabad are led by experienced instructors with extensive knowledge in the field. These instructors bring real-world expertise and industry insights to the virtual classroom, providing practical examples and guidance throughout the course.

  • Certification Opportunities:

Completing Linux Administration Online Classes may allow you to earn industry-recognized certifications. Choosing classes that align with recognized certification programs is essential to maximize the value of your learning journey.

Enhance Your Linux Expertise with RHCE, RHCSA, and Red Hat Training in Ahmedabad

In Ahmedabad, you can broaden your Linux administration skills through RHCE and RHCSA classes. These comprehensive programs offer a range of features to help you excel in Linux-based environments. RHCE RHCSA Classes in Ahmedabad provide in-depth knowledge and practical skills required to design, deploy, and manage Red Hat solutions effectively. Linux Training in Ahmedabad covers various topics such as system administration, network configuration, and security management. These institutions validate your expertise, enhancing your professional credibility. By enrolling in these programs, you can acquire valuable knowledge, hands-on experience, and potential career advancement opportunities in Linux administration.

Conclusion:

In a rapidly changing digital landscape, continuous education is vital for professionals seeking to stay ahead. Many offer a diverse range of online classes and training programs tailored to meet the demands of open-source technologies. By enrolling in Linux administration, Red Hat training, and certification courses, you can enhance your skill set and gain a competitive edge. Visit the highskyit.com website for more information and start your educational journey toward success.

29 June 2023

What Is Docker? How To Install RHEL 9

What Is Docker?

With the aid of containerization, you can deploy, scale, and manage apps automatically using the open-source platform Docker. To run an application reliably across various computer environments, containers are compact, isolated environments that package an application and its dependencies.

The following are some fundamental Dock ideas.

Images: To launch a containerized application, a basic operating system, application code, and dependencies must be defined in read-only templates called “docker images.” Images can be created from scratch or based on pre-existing images found in open or closed Docker registries.

Containers: Instances of Docker images are containers. They are easy to create, start, stop, move, and remove, and they are lightweight. Because each container is independent and has its filesystem, processes, and networking, they are reproducible and transferable between various settings.

Docker Compose: You may define and manage multi-container applications using Docker Compose. It makes it simple to create and run complicated, interconnected systems by using a YAML file to define the services, networks, and volumes needed by your application.

Dockerfile: You may define and manage multi-container applications using Docker Compose. It makes it simple to create and run complicated, interconnected systems by using a YAML file to define the services, networks, and volumes needed by your application.

Containerization: A container is created by Docker by encapsulating an application along with all of its dependencies, including libraries, frameworks, and tools. Consistency and mobility are ensured by the isolation of containers from one another and the underlying host system.

Docker Registry: A repository for Docker images is known as a Docker registry. The default public registry where you may get a variety of pre-built images is the Docker Hub. Additionally, private registries can be created so that you can share and keep your photographs within your company.

With Docker, applications can be packaged and distributed in a flexible and effective manner, resulting in shorter development cycles, increased scalability, and more effective resource use. Due to its simplicity of use, portability, and ecosystem of tools and services, it has grown to be a well-liked technology in the field of software development and deployment.

How To Install Docker :-
What is Docker? And How To Install In Ubuntu 20.04

Red Hat 9 is a fairly old version of Red Hat Linux, as was previously indicated, and Docker is incompatible with it. A contemporary operating system with a compatible kernel version is necessary for Docker.

Remove Old versions

The names of earlier versions of Docker were docker or docker-engine. Before attempting to install a new version and any associated dependencies, uninstall any such previous versions. Additionally, if Podman has already been installed, remove it and the related dependencies:

yum remove docker \
docker-client \
docker-client-latest \
docker-common \
docker-latest \
docker-latest-logrotate \
docker-logrotate \
docker-engine \
podman \
runc

All Posts

press y 

Complete ! Remove 

Installation methods Docker :-

Set up the repository and install the yum-utils package, which includes the yum-config-manager tool.

yum install -y yum-utils

yum-config-manager --add-repo https://download.docker.com/linux/centos/docker-ce.repo

Install Docker Engine

Installing container, the docker engine, and docker compose

yum install docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin -y

Start Docker.

systemctl start docker

of this command  ” enable”  automatically boot time. start  Docker service

systemctl enable docker

Check service running

systemctl status docker.service

28 June 2023

What is Docker? And How To Install In Ubuntu 20.04

Containers are standardized, executable components that integrate application source code with the operating system (OS) libraries and dependencies necessary to run that code in any environment. Docker is an open-source platform that empowers developers to build, distribute, operate, update, and manage containers.

The magic bullet that permanently fixed the virtualization and software container issues was Docker. Yes, that is a bold statement! Other products had made an effort to address these issues, but Docker’s novel strategy and ecosystem had completely eliminated the competition. You will learn the fundamentals of Docker in this course so that you can start utilizing it for your own applications and incorporating it into your workflow.

 1  Installing Docker

What Is Docker? How To Install Rehal 9

It’s possible that the Docker installation package included in the official Ubuntu repository is out of date. We’ll install Docker from the official Docker repository to make sure we have the most up-to-date version. To accomplish that, we will first create a new package source, then install the package after adding the GPG key from Docker to confirm the downloads are legitimate.

Update your current list of packages first:.

apt update

Install the following prerequisites to enable apt to use packages through HTTPS:

apt install apt-transport-https ca-certificates curl software-properties-common

curl -fsSL https://download.docker.com/linux/ubuntu/gpg | sudo apt-key add -

Then add your system’s GPG key for the official Docker repository:

APT sources should include the Docker repository:

add-apt-repository "deb [arch=amd64] https://download.docker.com/linux/ubuntu focal stable"

Additionally, this will add the Docker packages from the recently added repository to our package database.

Verify that you are about to install from the Docker repository rather than the standard Ubuntu repository:

The result will look like this, albeit Docker’s version number can be different:

Note that docker-cue is not installed, but Ubuntu 20.04 (focal)’s Docker repository is the installation candidate.

Install Docker lastly:

apt install docker-ce

Now that Docker has been set up, the daemon should be running and the process should be set to launch upon boot. Verify that it is operating:

To verify that the service is operational and operating, the output should resemble the following:

The Docker client as well as the Docker service (daemon) are now included with the installation of Docker. Later in this lesson, we’ll look at how to use the docker command.

Install docker in complete 

WhatsApp chat