This site uses cookies to improve your experience. To help us insure we adhere to various privacy regulations, please select your country/region of residence. If you do not select a country, we will assume you are from the United States. Select your Cookie Settings or view our Privacy Policy and Terms of Use.
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Used for the proper function of the website
Used for monitoring website traffic and interactions
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Strictly Necessary: Used for the proper function of the website
Performance/Analytics: Used for monitoring website traffic and interactions
We had some fun getting hardware figured out, and I used a 3D printer to make some cases, but the whole project was interrupted by the delivery of the iPhone by Apple in late 2007. Photo by Adrian2008 Netflix Hack Day Trophy for Instant Queue Add forNetflix In January 2010, exactly 15 years ago, Apple announced the iPad.
In my previous post , I reviewed historical data on single-core/single-thread memory bandwidth in multicore processors from Intel and AMD from 2010 to the present. The increase in single-core memory bandwidth has been rather slow overall, with Intel processors only showing about a 2x increase over the 13 years from 2010 to 2023.
This was a chance to talk about other things I've been working on, such as the present and future of hardware performance. The video is on [youtube]: The slides are on [slideshare] or as a [PDF]: I work on many areas of performance, but recently I've had a lot of demand to talk about BPF.
By Werner Vogels on 14 November 2010 04:00 PM. The early GPU systems were very vendor specific and mostly consisted of graphic operators implemented in hardware being able to operate on data streams in parallel. The Amazon.com 2010 Shareholder Letter Focusses on Technology. Comments (). General Purpose GPU programming.
We used this model effectively at Netflix when I was their cloud architect from 2010 through 2013. The layers of platforms start at the bottom with hardware choices such as which CPU architectures and vendors you want to use. There are three current underlying reasons for the platform engineering meme today.
By Werner Vogels on 12 July 2010 05:00 PM. Additionally, many high-end HPC applications take advantage of knowing their in-house hardware platforms to achieve major speedup by exploiting the specific processor architecture. The Amazon.com 2010 Shareholder Letter Focusses on Technology. All Things Distributed. Comments ().
Doubly so as hardware improved, eating away at the lower end of Hadoop-worthy work. ” There’s as much Keras, TensorFlow, and Torch today as there was Hadoop back in 2010-2012. Google goes a step further in offering compute instances with its specialized TPU hardware. Those algorithms packaged with scikit-learn?
Looking at sustained single-core bandwidth for a kernel composed of 100% reads, the trends for a large set of high-end AMD and Intel processors are shown in the figure below: So from 2010 to 2023, the sustainable single-core bandwidth increased by about 2x on Intel processors and about 5x on AMD processors.
Vivino was founded in 2010 and now has more than 20 million users worldwide and offices in four countries across three continents. If the solution works as envisioned, Telenor Connexion can easily deploy it to production and scale as needed without an investment in hardware. Our AWS Europe (Stockholm) Region is open for business now.
To reach those goals, Shell in 2010 began using AWS. When Tom Tom launched the LBS platform they wanted the ability to reach millions of developers all around the world without having them invest a lot of capital upfront in hardware and building expensive data centers so turned to the cloud.
Hardware gets better, sure. Let’s say just 2 billion people (somewhere around 4 billion are connected to the internet) view 5 pages at that weight in a day (surveys from 2010 showed the number to be around 10 sites on any given day, so I’m being overly safe here). It makes sense.
In my previous post , I reviewed historical data on single-core/single-thread memory bandwidth in multicore processors from Intel and AMD from 2010 to the present. The increase in single-core memory bandwidth has been rather slow overall, with Intel processors only showing about a 2x increase over the 13 years from 2010 to 2023.
I said something similar to the above, but with two important differences: I said hardware “threads,” not only hardware “cores” – it was about the amount of hardware parallelism available on a mainstream system. Figure 4 : How much concurrency does your program need in order to exploit given hardware?
Hardware Past As Performance Prologue. Regardless, the overall story for hardware progress remains grim, particularly when we recall how long device replacement cycles are: Tap for a larger version. But the hardware future is not evenly distributed, and web workloads aren't heavily parallel. Mind The Gap.
This was a chance to talk about other things I've been working on, such as the present and future of hardware performance. The video is on [youtube]: The slides are [here] or as a [PDF]: first prev next last / permalink/zoom I work on many areas of performance, but recently I've had a lot of demand to talk about BPF.
After the release of the iPad in 2010 Craig Hockenberry discussed the great value of communal computing but also the concerns : “When you pass it around, you’re giving everyone who touches it the opportunity to mess with your private life, whether intentionally or not. This expectation isn’t a new one either. Lots of problems, now what?
In both cases, the OS will task the browser team to heavily prioritise integrations with the latest OS and hardware features at the expense of more broadly useful capabilities — e.g. shipping "notch" CSS and "force touch" events while neglecting Push. Examples include IE 7+ and Safari from 2010-onward.
As for attending USENIX conferences: I finally started attending and speaking at them in 2010 when a community manager encouraged me to (thanks Deirdre Straughan), and since then I've met many friends and connections, including Amy who is now USENIX President, and Rikki with whom I co-chaired the USENIX LISA18 conference.
The success of both tablets and smartphones (and to a lesser extent the fringe battlegrounds of e-readers and other specialized devices), as well as the rapid maturation of cloud-based services, has created a tech hardware war, an OS war, a bidding war for tech firms, and spawned a feeding frenzy in application development.
Looking at sustained single-core bandwidth for a kernel composed of 100% reads, the trends for a large set of high-end AMD and Intel processors are shown in the figure below: So from 2010 to 2023, the sustainable single-core bandwidth increased by about 2x on Intel processors and about 5x on AMD processors.
The thing about any crisis is that there's an expectation that there will eventually be a recovery, no matter long it takes to reach bottom and regardless the shape the recovery takes: V-shaped rebound as happened in in 1953, W-shaped as in 1981-82, U-shaped as in 1973-75 and again in 2000-03, or L-shaped as in 2008-2010.
HTML, CSS, images, and fonts can all be parsed and run at near wire speeds on low-end hardware, but JavaScript is at least three times more expensive, byte-for-byte. Since 2010, volumes have been on a slow downward glide path , shrinking from ~350MM per year in a decade ago to ~260MM in 2018.
In 2010, Netflix introduced a technology to switch production software instances off at random — like setting a monkey loose in a server room — to test how the cloud handled its services. If any touchpoint in the cloud failed, the quality of the viewers’ experience could degrade.
The webp format was introduced by Google around 2010. There is a potential benefit in reusing the hardware in place for video compression/decompression. With the motion extension, it was accepted as the video coding standard for digital cinema in 2004.
Around 2010, when US companies could count on a recovering domestic economy, a weak US dollar, and growing emerging market economies, CFOs didn't need to work that hard to flatter the income statement. We know that this echoes the supercycle of businesses separating into asset-light operating companies (e.g.,
We organize all of the trending information in your field so you don't have to. Join 5,000+ users and stay up to date on the latest articles your peers are reading.
You know about us, now we want to get to know you!
Let's personalize your content
Let's get even more personalized
We recognize your account from another site in our network, please click 'Send Email' below to continue with verifying your account and setting a password.
Let's personalize your content