Goal of this blog is, understanding what DevOps means, how teams implemented or experienced DevOps in 2018 and what is in store for 2019. The biggest issue while adopting any technology or framework is, creating own definitions or using them for the purposes which those are NOT designed for. Let us avoid this trap by doing a quick review on what DevOps and how it works, this will help us understand the blog content very well.
DevOps is combination of practice, tools and philosophy. Teams who adopt DevOps consists of complimentary skilled engineers who work across the entire application lifecycle. For example, people who are skilled in software development, testing, deployment and operations do develop a range of skills not limited to a single function.
NVTC’s Tech 100 list includes some of the region’s top companies & individuals driving tech innovation in areas like Cloud, IoT, Big Data and Analytics etc., and leading growth in the Greater Washington Metropolitan area. Companies and individuals were nominated for consideration by independent parties and selected as winners by a judging committee.
AWS EC2 instances are one of the widely used resources available in AWS ecosystem. Amazon Elastic Compute Cloud (Amazon EC2 ) is a web service that provides secure, resizable compute capacity in the cloud. It is designed to make web-scale cloud computing easier for developers. Out of all types of EC2 instances, behaviour of T2 type instances are little bit tricky in terms of system functionality. When users launch T2 (standard) type of servers, AWS imposes certain restriction on available CPU. Because T2 types are supported in free tier and widely used by new users, understanding how this CPU credit allocations works and impacts your EC2 instance performance will make AWS users life easy in deploying correct type of applications on these instances during testing.
Cloud runs on per second billing. When companies rent servers from AWS, they will get billed for every second (except for windows OS servers) they keep their EC2 instances up and running irrespective whether anyone access the applications hosted on them or not. AWS billing engine doesn’t care whether EC2 instance is being utilized 100% or idle. If it is in “running” state, it will be charged for compute hours. Dictionary definition of a “schedule” is: Arrange or plan (an event) to take place at a particular time Last two words in the definition are very crucial. Top most issue with schedulers is, identifying “particular time”, in other words “schedule”. . In dynamic and agile world, users do multitasking and access servers or applications on-demand need basis. Associating a schedule to anything which is need basis would severely limit the capabilities and effect outcomes.
AWS Virtual Private Cloud (VPC) is now the default scheme for running cloud VMs.Your VPC can resemble a traditional on-premises network but with more automation and scale.Amazon Virtual Private Cloud (VPC) is the heart of AWS cloud hosting, yet a very complex concept to understand, especially for developers who have limited infrastructure operations experience. Developers are the most involved team members with cloud projects, yet have limited knowledge about infrastructure operations ((in the majority of cases).
originally published on Infoq
In a typical company using AWS EC2 for their cloud hosting, there will be different environments like production code running on few AWS EC2 instances, few others running QA tests, few more EC2 instances for DEV activity etc., These possible different deployment scenarios require different operating needs is the primary reason why we advise “one cost optimization solution” will not yield best possible optimization in cloud savings. Combination of “Right Capacity Sizing” and “EC2 Reserved Instances (RI)” will help companies reduce waste on AWS EC2 spending. But, can these same principles be applied to other deployment environments like DEV, QA servers etc.,? BIG NO. WHY?
"On-premise vs Cloud", this is one of the first question every company/team who are exploring public could adoption would ask. Available answers over internet for this question at this point are either dedicated to one topic (or) limited to few tips based on authors experience. Companies are in real need of a repository with this sort of information to make an informed decision about their cloud adoption. I am creating this page as repository to pool the information from experts. I feel that repository like this will tremendously benefit everyone. Following are the topics I have in my mind. Feel free to suggest topics and respective differences in On-premise vs Cloud practices.
Recent RightScale study observed that around $10B was wasted on cloud spend and one of the top reason for this waste is leaving ec2 instances in running state when no one using them. We tried solutions like ec2 schedulers. But, they are not optimal solutions to save on AWS instance costs. Why? Let us quickly check it. In cloud computing era, where most of the resources are billed per second, leaving servers up & running when no one using them, will not help much with your AWS cloud bill and not best optimization.
f you are a website owner who bought a domain from GoDaddy (or some other registrar) and started exploring AWS to take advantage of the capabilities of cloud computing, one of the things you need to address will be: my application/website servers are hosted on AWS, so how can I integrate with the domain I bought from GoDaddy? The easiest answer is to update your GoDaddy NS records to point to Amazon’s name servers. In this tutorial, we are going to look at how this setup works. Note that we’ll be looking at GoDaddy so that we have a specific example, but this process will be very similar no matter where you bought your domain name.
It is also an ideal platform for organizations that have compliance concerns. INVOKE Cloud takes DevOps to next level by making it TeamOps. Everyone with the permissions is able to bring up servers on-demand, not just developers. INVOKE Cloud lets users bring up their cloud servers whenever they are down by simply typing in the application URL from anywhere using any browser. The software also provides configuration options to filter users and groups who can bring up the servers and applications on-demand.
Our guest blog on DevOps.com website.
In cloud DevOps, however, that may not be the case. Any cost-conscious organization may instruct developers to take infrastructure offline when it’s not needed (most of us know how developers manage infrastructure, so I am not going discuss much about it). But what about other stakeholders of the project? What if the QA team wants to manually test or validate something or product owners want to review something?
Most of us know that 3 key pieces of modern computing are: CPU, Memory (RAM & Hard Disk) and I/O. All cloud hosting providers price the usage of these key components (either hourly (or) minute (or) seconds based on provider). Let us explore how & what AWS price the usage of these components, for their cloud consumer.
From the moment you launch an instance until termination, EC2 instances transitions between different states. The transitions are very clear and easier to understand, we are not going to discuss about state transitions here. What we are going to discuss is, how billing will get affected when you move EC2 instances between these states.
READY TO SAVE ON CLOUD COSTS? Get Started for free