This site uses cookies to improve your experience. To help us insure we adhere to various privacy regulations, please select your country/region of residence. If you do not select a country, we will assume you are from the United States. Select your Cookie Settings or view our Privacy Policy and Terms of Use.
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Used for the proper function of the website
Used for monitoring website traffic and interactions
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Strictly Necessary: Used for the proper function of the website
Performance/Analytics: Used for monitoring website traffic and interactions
On Tuesday, January 3, 2023, at 15:26 UTC, we experienced an interruption of Dynatrace’s Single Sign On (SSO) service, preventing our customers from logging into their Dynatrace Software as a Service (SaaS) environments and other Dynatrace portals. Our customer’s monitoring data collection was not affected, apart from a few identified and informed customers with a larger dependency on SSO authorization.
Last time I blogged about the New WAL Archive Module/Library feature available in PostgreSQL 15 , which is quite transformative in how WALs are archived today in PostgreSQL. PostgreSQL 15 has many more improvements related to WAL archiving, which is worth discussing. In this blog, I would like to highlight some of them which solve great operational challenges for many of the PostgreSQL users.
This article is meant to be a starting point for sharing observations, experiences, and ideas to begin and build a conversation, exchange experiences, discuss, and shape a product vision we can work towards as a community.
As we close out 2022, I thought I’d write a short update on what’s been happening in Cpp2 and cppfront. If you don’t know what this personal project is, please see the CppCon 2022 talk on YouTube. Most of this post is about improvements I’ve been making and merging over the year-end holidays, and an increasing number of contributions from others via pull requests to the cppfront repo and in companion projects.
We have released Dynatrace version 1.257. To learn what’s new, have a look at the release notes. The post Dynatrace SaaS release notes version 1.257 appeared first on Dynatrace news.
In PostgreSQL, vacuuming is a maintenance task that helps to optimize database performance and reclaim space. It involves removing deleted or outdated rows from tables and indexes and updating statistics used by the query planner. This process is necessary to prevent the accumulation of unnecessary data, known as “dead rows,” which can take up significant space and slow down queries.
Have you ever needed to handle signal events within your ZX script? For example, what happens if you need to handle a graceful shutdown of your long-running or infinite-running script? Or what happens when the user forcefully stops the script? These signal events are typically “SIGTERM”, “SIGINT”, and similar events. When using a script language like Bash, these events are most commonly captured with trap commands, but what happens when we’re using ZX?
Percona’s XtraBackup is a beautiful tool that allows for the backup and restoration of MySQL databases. From the documentation: The Percona XtraBackup tools provide a method of performing a hot backup of your MySQL data while the system is running. Percona XtraBackup is a free, online, open source, complete database backups solution for all versions of Percona Server for MySQL and MySQL®.
Sign up to get articles personalized to your interests!
Technology Performance Pulse brings together the best content for technology performance professionals from the widest variety of industry thought leaders.
Percona’s XtraBackup is a beautiful tool that allows for the backup and restoration of MySQL databases. From the documentation: The Percona XtraBackup tools provide a method of performing a hot backup of your MySQL data while the system is running. Percona XtraBackup is a free, online, open source, complete database backups solution for all versions of Percona Server for MySQL and MySQL®.
As the globe strides into 2023 — with rapid change and macroeconomic uncertainty looming — organizations want tools and technologies that enable them to become more efficient, reduce costs, and innovate more. These are precisely the business goals of AIOps: an IT approach that applies artificial intelligence (AI) to IT operations, bringing process efficiencies.
MySQL index is a data structure used to optimize the performance of database queries at the expense of additional writes and storage space to keep the index data structure up to date. It is used to quickly locate data without having to search every row in a table. Indexes can be created using one or more columns of a table, and each index is given a name.
Part Three of the Humanizing Software Quality Series from Intellyx, for Apica. In part 1 of this series, Jason English established user journeys as the essential goal of software testing efforts. Then in part 2 , he explained why user journeys are so difficult to test. The fundamental lesson across both articles is that the software quality buck stops with the user.
FIFA World Cup 2022 was a massive spectacle in viewership and marketing revenue. Unfortunately, such an enormous scale of events caused much chaos for some of the apps streaming this global event. This article analyzes those apps from a performance engineering point of view, highlights the reasons for their crash or lags, and tries to provide possible solutions.
In Managed Services , we have many customers, and as each has a different kind of config and environment, working on their environment is always fun and interesting. In this blog post, I will showcase an issue we faced when dropping a table and how it was resolved. There was a ticket to drop a table in a client’s production environment (MySQL 5.7).
Southwest Airlines has made headlines in recent days for all the wrong reasons: bad weather impacted air travel, which required Southwest to adjust plane and crew schedules. Those adjusted schedules were often logistically flawed because the planes and crews matched at a specific place and time didn’t make sense in the real world. Making matters worse, those adjusted schedules had to be re-(and re- and re-)adjusted every time either the weather changed or operations changed (ie., more flight can
Recently I was working with a customer wherein our focus was to carry out a performance audit of their multiple MySQL database nodes. We started looking into the stats of the performance schema. While working, the customer raised two interesting questions: how can he make complete use of the performance schema, and how can he find what he requires? I realized that it is important to understand the insights of the performance schema and how we can make effective use of it.
Space constraint has been an endless and painstaking process for many of us, especially in systems that have a high number of transactions and data growth exceeding hundreds of GBs in a matter of days. In this blog, I will share a solution to remove this space and remove rows from tables in a few seconds regardless of the size of a table without causing any additional load on the database using table partitions.
Recently, we performed a database engine major version upgrade in one of our customers’ environments from MySQL 5.7.26 to 8.0.27. After this version upgrade, we experienced issues with backups and replication for one of the nodes. In this article, I will explain these issues in detail and recommend a way to take backups from a replication environment.
Moving to the cloud is one thing, and saving costs is another. One of the reasons why organizations migrate to the cloud is to reduce infrastructure costs. However, caveats like overprovisioning, wasted resources and management issues lead to higher cloud costs. So, here are some best practices to reduce cloud costs and save more on your bills!
Fine-tuning your cloud infrastructure is critical to ensure your overall bill keeps up to its limit. Read this blog to find out proven best practices for cloud cost optimization to help you cut down on the bill and save costs by eliminating unused resources.
We organize all of the trending information in your field so you don't have to. Join 5,000+ users and stay up to date on the latest articles your peers are reading.
You know about us, now we want to get to know you!
Let's personalize your content
Let's get even more personalized
We recognize your account from another site in our network, please click 'Send Email' below to continue with verifying your account and setting a password.
Let's personalize your content