Kalypso has been fighting the good fight recently, ensuring forgotten PC gaming genres don't stay that way for good. Railway Empire's inspirations are immediately obvious to Railroad Tycoon fans, promising plenty of tracklaying fun as would-be tycoons attempt to appease both the passenger and service markets simultaneously. For those types like myself who love a bit of tycooning it's looking awesome, but as always, be sure to check out the official Railway Empire PC system requirements first.
As far as system requirements go, these are right down at the lower end of the spectrum. Anyone with a PC bought in the last three or four years should have literally problems to worry about in terms of running Railway Empire. The only conceivable issues arise if you're still running any very old components in your gaming PC.
It's been a full decade since we've been treated to a SpellForce sequel. An absolute cult favourite, Grimlore Games knows how to keep its fans waiting. SpellForce 3 will be bringing its unique blend of real-time strategy and role-playing to PC next month though, and it's looking like a very intoxicating combo indeed. As far as strategy games go, Spellforce 3 looks visually spectacular, so here are the system requirements you'll need if you hope to run Spellforce 3 on PC.
In the grand scheme of things, these system requirements actually hover right around the average system requirements for a AAA game in 2017. It's undoubtedly the CPUs which could probe the most problematic for those looking to run Spellforce 3 though. Even the minimum spec Intel Core i5-3570 is still decent by today's standards so you'll be wanting a moderately capable quad-core processor if you want to run Spellforce 3 without a hitch.
As if the rising price of memory and spiralling demand for crypto-mining GPUs wasn’t enough, the cost of silicon wafer is set to increase dramatically over the next few years. It’s time to start paying wallet emptying prices for glorified sand.
SUMCO is a Japanese silicon wafer manufacturer estimated to provide two-thirds of the world’s supply of silicon wafer. This silicon wafer is one of the raw materials used in semiconductor manufacturing, used in everything from CPUs to GPUs and DRAM. SUMCO has confirmed the price of silicon wafer has jumped 20% this year alone, and it’s set to continue rising substantially for the next few years at least.
Samsung’s confirmed its upcoming GDDR6 memory will be faster than previously anticipated, clocking in at 16Gb/s data rate. Originally, when Samsung revealed its GDDR6 memory last year, it was expected to top out at 14 Gbps. yet now, as it enters mass production, Samsung has bumped up its own targets to 16 Gb/s.
“The fastest and lowest-power DRAM for next generation, graphics-intensive applications,” teases Samsung in a new statement. “It processes images and video at 16Gbps with 64GB/s data I/O bandwidth, which is equivalent to transferring approximately 12 full-HD DVDs (5GB equivalent) per second. The new DRAM can operate at 1.35 volts, offering further advantages over today’s graphics memory that uses 1.5V at only 8Gbps.” I have to say that little DVD fact is kind of blowing my mind. Now all I can do is look at my USB cables and scowl. It should mean 4K bandwidth is a breeze for GPUs which utilise it.
Forget Volta - the latest rumours surrounding Nvidia’s next-generation GPU architecture suggests it will, in fact, be GeForce Ampere. According to Heise.de, Nvidia is planning to reveal/pre-launch Ampere during GPU Technology Conference in March, much earlier than the second half of 2018 that we were expecting.
The Ampere microarchitecture may be an evolution of the Volta chips used in Nvidia’s supercomputer GPUs (or a devolution without Tensor Cores), or it could refer to cut down consumer versions designed to bring affordability to the next-generation. Whatever the case, the rumours suggest Ampere will be the successor to Nvidia’s GeForce 10 series graphics cards. I think we’re looking at either GeForce 11 or GeForce 20.
The first photo has emerged of Intel and AMD’s joint venture. The MCM (multi-chip module) chip in question is an Intel CPU with an AMD Radeon GPU onboard, along with a 4GB stack of HBM2 memory.
Nvidia’s just hosted its quarterly earnings call and questions naturally arose regarding AMD Radeon boss Raja Koduri’s defection to Intel, heading up development of discrete graphics cards.
Unsurprisingly, Nvidia CEO Jen-Hsun Huang said AMD losing Koduri to Intel was a big loss, causing a significant drain in know-how from one of AMD’s most recognisable figures.
Intel has appointed ex-Radeon Technologies Group boss Raja Koduri, confirming the news rumoured earlier this week. In another surprising twist, Koduri will be joining Intel as chief architect and senior vice president of its new Core and Visual Computing Group. He’ll be working on high-end graphics card solutions and going head-to-head with AMD Radeon and Nvidia GeForce. It’s not often that the hardware industry turns into a full-blown soap opera, but it is today.
"Raja is one of the most experienced, innovative and respected graphics and system architecture visionaries in the industry and the latest example of top technical talent to join Intel,” said Dr. Murthy Renduchintala, Intel's chief engineering officer and group president of the Client and Internet of Things Businesses and System Architecture. Try fitting that on a business card.
Radeon Technologies Group head honcho Raja Koduri has officially left AMD. The confirmation comes from the man himself, who wrote a heartfelt letter to the AMD faithful explaining his departure. It was roughly two months ago when Koduri announced he would be taking a three-month sabbatical from AMD for personal reasons.
“Forty is a significant number in history," writes Koduri. "It is a number representing transition, testing and change. I have just spent forty days away from the office going through such a transition. It was an important time with my family, and it also offered me a rare space for reflection. During this time I have come to the extremely difficult conclusion that it is time for me to leave RTG and AMD.
After months of rumour-mongering, Intel has at long last come out and officially announced a deal with AMD to include Radeon Technologies Group semi-custom graphics chips on its Kaby Lake G-series CPUs.
The deal incorporates a discrete AMD Radeon GPU on a high-performance Intel CPU, including 4GB of onboard HBM memory. The result is an ultra-thin APU with fast compute performance and a low power draw. It’s all stitched together with a custom interconnect, a result of Intel's intent to move towards heterogeneous architecture.