Jump to content

AMD announce new processors


Ray Proudfoot

Recommended Posts

Ray Proudfoot

Overnight AMD have released news on their latest Ryzen 7000 processors. Clock speed up to 5.7Ghz. B)

In late September Intel will be announcing their latest processors.

I was going to start deciding on a new PC this month but hanging on a little longer seems sensible.

Link to comment
Daniel Saffran

I've gone for AMD on my current Build to max out on Cores and VRAM and learned one Thing: I'll go for Intel+Nvidia again (Rocket-Lake / Lovelace) as soon as they are out. One Reason being that Prepar3D is just not able to use the Resources efficiently. The other Reason being stability and annoying Issues I never have experienced on Intel (USB Discos, Microstutters ...).

"Disclaimer"
I don't want to start a "Team Blue/Green or Red is better / the best" Discussion / War. We live in times where you can again get great (and bad!) Products from either Party (if available xD). So it is even more personal choice now, and that is great!
I just want to share my personal Experience about that one time trying AMD. Never really regret it, but would not do it again either :) 

Link to comment
Rob Ainscough

Cool.

Intel will be releasing Raptor lake 13900K Oct 13th I believe (with pre-orders on Sept 28th one day after AMD makes 7000 series available for sale), or at least making some type of announcement ... it is supposed to be a tad faster in single thread performance by 4% over the7950X and has more cores 24 (but keep in mind eCores).

Intel's 13000 series, the13900K is the only one that outperforms AMD 7950X ... however, AMD have yet to implement 3D V-cache for the 7000 series CPUs which should be available Dec 2022 or Jan 2023.

Intel CPU overclock process is very different, more a brute force approach using lots of POWER, must have MORE POWER (in my best Arnold S. voice).  Intel still 7nm process, AMD is 5nm process.

AMD have F-Max which limits the boost frequency to 5.85 Ghz no matter what one does (you have to get very creative if you want to get around that, think major hacks).

Either way, I'd wait a month or two ... the 7900X is probably the best performance bang for the buck even after Intel's release.

  • Like 2
Link to comment
Rob Ainscough
26 minutes ago, Daniel Saffran said:

The other Reason being stability and annoying Issues I never have experienced on Intel (USB Discos, Microstutters ...)

There was an EFI/BIOS update for AMD 5000 series that solved the micro stutter issue ... it was related to the TPM (Trusted Platform Module).  Intel have their own fair share of little issues from performance to security to power to heat.  Can't say I've noticed any difference in stability between the two after OC testing and getting Intel/AMD working optimally ... both seem very stable.

AMD responds to lower latency RAM better than Intel, so RAM choice is very important with AMD, I gained a 15% increase across the board by simply moving up from 3000Mhz CL16 64GB to 3600Mhz CL14 64GB.  

BUT Microsoft clearly favor Intel and have added OS support directly for Intel and not for AMD ... so AMD does have an uphill battle when both OS vendor (MS) and CPU vendor (Intel) are working very closely together.  Something to think about.

On a side note, if you do any sort of benchmark testing, be sure to disable Process Lasso (or similar tools) otherwise you'll probably end up with lower performance scores.  Process Lasso works well for P3D in terms of multiple core efficiency but disable it if you do any benchmarking.

I'll be buying a 7950X Sept 27th (I hope, if I can beat the BOTs) along with some good DDR5 RAM ... plan to pair it up with AMD's RNA3 GPU later this year as the WATT requirements for nVidia 4000 series and out of this world price has made it more affordable to fly a mission to MARS in the real world. ;)  

Cheers, Rob.

  • Like 2
Link to comment
Markus Burkhard
4 hours ago, Rob Ainscough said:

Intel will be releasing Raptor lake 13900K Oct 13th I believe (with pre-orders on Sept 28th one day after AMD makes 7000 series available for sale), or at least making some type of announcement ... it is supposed to be a tad faster in single thread performance by 4% over the7950X and has more cores 24 (but keep in mind eCores).

Sure, release the 24 cores monster first, the one that gamers don't need but will buy anyway, because they'll want it immediately. Bit of a cash-grab there by Intel.

Buying a 24 core CPU for flightsim purposes is complete nonsense. Even though both P3D and MSFS can use multiple cores, and so does FSL software, nothing in PC flightsim will be optimised to really take advantage of that many cores, for many years to come. Heck DCS is still basically a single-core sim. 
So I would recommend anyone reading not to fall for the 13900K trap. Save money (and power requirements) by either shopping for the new AMD 7000 series, or at the very least wait for a 13000 series Intel CPU that is at least one model below the flagship.
For me personally the AMD 7000 series seems to be the winner this time around, as the rumoured power requirements of the next gen Intel silicon is borderline insanity. I'll hold my final judgment for when the specifications are here, but looking at the 12000 series chips and the rumours for the new one, it does not look too promising for Intel to be perfectly honest. 

  • Like 5
Link to comment
Rob Ainscough
6 minutes ago, Markus Burkhard said:

Even though both P3D and MSFS can use multiple cores

That's false, well at least for P3D ... skip forward to 10:27 and you can see plenty of multi-core activity:

Agree that 24 cores would not be terribly useful for P3D, 12 real cores would do it for me.

Cheers, Rob.

 

  • Like 1
Link to comment
Markus Burkhard
11 minutes ago, Rob Ainscough said:

That's false, well at least for P3D ... skip forward to 10:27 and you can see plenty of multi-core activity:

Agree that 24 cores would not be terribly useful for P3D, 12 real cores would do it for me.

I know Rob, that's what I was trying to say. It can use multiple cores, and yes will of course also "use" 24 of those. But it is nowhere near optimised for that many. Hence me recommending to save money there and use it for a better GPU, faster RAM, fast SSD and so on.

Link to comment
Rob Ainscough

Agree that 24 cores will just add overhead to the main thread ... for my needs 12 "real cores" (not eCores, not SMT, not HT) cores would be the sweet spot ... which falls right into AMD 7900X spec ($549).  AMD has 76-80MB Cache, Intel's 68MB Cache ... there might be an advantage to disabling the cores on the 13900K as the cache is shared.

Huge benefit to all is the competition between AMD/Intel will help keep prices lower.  I also hope AMD's next gen GPU brings enough grunt to the table to force nVidia to rethink it's insane "projected" MSRP for the 4000 series.  We really NEED AMD as viable competition.

  • Like 3
Link to comment
Nuno M Pinto

I also agree that the 7900X seems to be the best next-gen CPU to go for. I'm not in the hunt for a new system, but if i was i would definitely go for that one.

Link to comment
Daniel Saffran
19 hours ago, Rob Ainscough said:

There was an EFI/BIOS update for AMD 5000 series that solved the micro stutter issue ... it was related to the TPM (Trusted Platform Module).  Intel have their own fair share of little issues from performance to security to power to heat.  Can't say I've noticed any difference in stability between the two after OC testing and getting Intel/AMD working optimally ... both seem very stable.

Yeah I know, always updated to the latest BIOS (and therefore AGESA). But it took very long / multiple Version until AMD got it sorted out (mostly, still have some micros now).

Sure, had / have now fundamental stability Issues either, even OCed! :) 
It was just Sim-wise unstable when you loose your Joysticks and it was annoying to regularly hear the crippled sound due to stutters.

19 hours ago, Rob Ainscough said:

AMD responds to lower latency RAM better than Intel, so RAM choice is very important with AMD, I gained a 15% increase across the board by simply moving up from 3000Mhz CL16 64GB to 3600Mhz CL14 64GB.  

I would add that it is more important to use at least 3600 MHz to have the FCLK on 1:1 (and at that speed the lowest CL which one can afford). For Ryzen 5xxx - has to be seen what the Sweet-Spot for 7xxx is for a 1:1 Clock-Ratio!

Link to comment
Rob Ainscough

Yeah it did take a while for AMD to get it sorted.  I've never had any USB issues (47 USB devices attached) but I was in late on the AMD 5950X bandwagon as I refused to pay over MSRP so waited long after release date.

I can keep my AMD 5950X very cool, considerably cooler than any Intel CPU I have.  Even chilled my Intel CPU would frequently hit 90 C under heavy loads, my AMD 5950X hits 56 C under very heavy loads.  Intel was MUCH harder to OC.

 

Link to comment
Daniel Saffran

woah?! 56 C for an OCed 5950x? Impressive, what do you use as Cooler? I use mine as Room-Heater :lol:

 

What I forgot to add for the Reasons moving back to Intel/Nvidia: The "incompatibilities" between P3D and AMD GPUs (white flickering on SSAA >= 4, grid pattern on transparent Textures). Though they are purely LM's fault, there is no real hope that they will fix it.

Link to comment
Markus Burkhard
50 minutes ago, Daniel Saffran said:

What I forgot to add for the Reasons moving back to Intel/Nvidia: The "incompatibilities" between P3D and AMD GPUs (white flickering on SSAA >= 4, grid pattern on transparent Textures). Though they are purely LM's fault, there is no real hope that they will fix it.

I guess you're aware of this but I would just like to add for anyone else reading here; There is no requirement to pair Intel CPUs with Nvidia GPUs, there's no mandatory link between any of AMD's, Intel's or Nvidia's products.
So if someone has issues with AMD GPUs in a sim, there's no reason to switch the entire platform to Intel if a Nvidia GPU is desired. Nor does one need an AMD GPU if one of the new Ryzen 7000 series CPUs is desired. 

There is just a small advantage if one uses both AMD CPU and GPU, in which there is certain memory sharing capability available that might increase performance in certain use-cases. But I've never read anything about this being a groundbreaking advantage. It's more like an added bonus but does not necessitate to pair the two.

Anyone is free to match between the three suppliers this autumn. I would just like to remind that it will be important to not simply compare price vs. performance, but also take power consumption into consideration. 

  • Like 3
  • Thanks 1
Link to comment
Rob Ainscough
4 hours ago, Daniel Saffran said:

woah?! 56 C for an OCed 5950x?

That's my peak under heavy loads during benchmark, not my average. :)  ... I built my silent chiller to keep my main FS PC cool and quiet.

 

 

 

I'm working on another project that'll make this one look like child's play ... stay tuned.

FYI, I have Tinnitus (since about age 21) hence why my push to go sub-ambient and do it relatively quietly as constant frequency noise makes life uncomfortable for me, that was "one" of a couple motivations for this projects.

My SimPit, not setup with A2A, looks better with an FSL aircraft, more displays and buttons active :)

 

Cheers, Rob.

 

  • Like 1
  • Thanks 1
Link to comment
Daniel Saffran
On 9/4/2022 at 3:20 PM, Markus Burkhard said:

I guess you're aware of this but I would just like to add for anyone else reading here; There is no requirement to pair Intel CPUs with Nvidia GPUs, there's no mandatory link between any of AMD's, Intel's or Nvidia's products.

Yeah, absolutely true! Thanks for Clarification ;) 

Link to comment
Rob Ainscough

Given nVidia's recent announcement they plan to manipulate the 4000 series prices due to excessive 3000 series inventory ... I guess they didn't think Crypto would fail?? 

This puts AMD CPU and GPUs (RNA3 7000 series) in a great position.  How many ways can Jensen try to screw non-crypto customers ... all those direct deals to crypto miners and now holding back 4000 series supply so as to offload old inventory.  Not sure making this public at a shareholders meeting was such a great idea nVidia?

 

  • Like 1
Link to comment
  • 2 weeks later...
Rob Ainscough

So now EVGA is not longer a nVidia partner (EVGA stopping all GPU production) ... that's  a shame, I only buy EVGA PSUs and GPUs.  Cements my purchase of AMD going forward.  Design and competition are vital to a good consumer price to performance ratio ... nVidia play on being like Apple.

Link to comment
  • 2 weeks later...
Rob Ainscough

Watched a few actual real world performance testing videos ... ouch for Intel ... AMD 7950X ordered and on it's way ... 

uc?export=view&id=1kSXSRYLri3mdupaOdHBhf

Now I need to find a motherboard and some quality DDR5 RAM ... 

Link to comment
Ray Proudfoot

$740 will equate to around 700GBP given the way Sterling is going. We’re you not tempted to wait for Intel with 6Ghz rumoured?

Link to comment
Rob Ainscough

AMD 7950X is $699 ... Oregon so no tax, $40+ is for expedited shipping.

Intel can't run boost on more than 2 cores and power consumption is much higher, AMD 7950X is only about 10 watts higher than my 5950X.  

I'm also very interested in DE lidding results.  But with that said, I'll be in the market for another CPU for my Development PC so if the 13900K pans out I'll cover my bases.

But VERY unlikely I'll get nVidia 4000 series GPU anything, not at $1599 and 660 watts.

I remember when BPS was $2.06 dollar to 1 BPS ... 10p would actually by lots of sweets from the tuck shop. :)

Link to comment
Rob Ainscough

FYI, ASUS extreme motherboard is $1000, DDR5 6600 Mhz RAM CL32 64GB $700, EK AM5 CPU water block $180 ... wait on the water block, due Oct 21.

Only 3 PCIe slots, ugh and spacing between my GPU fully enclosed (front/back) water block (3090) is too close to my RAM water block so can't use PCIe slot one.

My issue is that I need to run a 2nd PCIe US card for all my other FS devices.

Waiting to build this out before I contact Pete so I can give him some metrics on performance with his P3D setup and see how/if he wants to proceed.

 

  • Like 1
Link to comment
Rob Ainscough

uc?export=view&id=1k1m04BID_Ll_YDiNzEU2C

Arrived

  • Like 1
Link to comment
Alexander Polcher

@Markus Burkhard regarding the current discussion I would like to enhance the topic and ask you about your opinion regarding a "minor upgrade".

My system specs as shown below (CPU i7 8700k @ 5,0GHz with AIO) have changed as I have ordered a new PSU (be quiet! 750W Pure Power 11 FM Modular 80+ Gold) and I am searching for a new GPU. The current situation shows some opportunities for RTX 3080 (especially Founders Edition) and RTX 3080 Ti. The 3080 FE would be available on eBay for 600€ incl. OVP plus receipt and 3080 Ti's are around 800-850€. 3080's are usually equipped with 10GB RAM and 3080 Ti's with 12GB.

Will the Ti version anyhow be bottlenecked or slowed down by my i7 8700k and a 3080 would be sufficient, or would the Ti version bringt a noticeable performance boost?


Would be nice to have some hardware crack feedback because I am literally lost in this topic and the new 40xx series seems not to be suitable for me as I do not own my own powerplant :D

Link to comment
Rob Ainscough

May I suggest solar panels and a good battery storage system, save you considerable money in the long run. :) 

Is this for MSFS or P3D or both? 

One of my test PC's has a 8700K (stock clock), I only saw 5% FPS improvement going from a 1080Ti to a Titan RTX (2000 series).  On my 9900K I saw more of an improvement around 19% when going from 1080Ti to Titan RTX (2000 series).  In this case, testing was done with P3D running 1440p res and 4XSSAA.  On my AMD 5950X and nv 3090, I got about 35% FPS improvement and able to run higher GPU intensive settings (like DR, more shadows, higher res clouds).

My advice, if you have 850E to spend and this is for P3D and you have at least a 1080Ti currently, then replace the CPU  with AMD 7000 series get one of the cheaper AMD motherboards and 32GB DDR5 6000Mhz RAM.  If this is for MSFS, then keep the CPU and go with 3080 GPU ... but be careful with eBay as there is a flood of GPUs that have been running 24/7 doing crypto mining.

EDIT: if you go with AMD, make sure you get good RAM ... performance difference can be as much as 19% just by RAM alone.  Intel seems to be less impacted by RAM speeds/latency.

Cheers, Rob.

  • Like 2
Link to comment
Alexander Polcher
16 minutes ago, Rob Ainscough said:

May I suggest solar panels and a good battery storage system, save you considerable money in the long run. :) 

Is this for MSFS or P3D or both? 

One of my test PC's has a 8700K (stock clock), I only saw 5% FPS improvement going from a 1080Ti to a Titan RTX (2000 series).  On my 9900K I saw more of an improvement around 19% when going from 1080Ti to Titan RTX (2000 series).  In this case, testing was done with P3D running 1440p res and 4XSSAA.  On my AMD 5950X and nv 3090, I got about 35% FPS improvement and able to run higher GPU intensive settings (like DR, more shadows, higher res clouds).

My advice, if you have 850E to spend and this is for P3D and you have at least a 1080Ti currently, then replace the CPU  with AMD 7000 series get one of the cheaper AMD motherboards and 32GB DDR5 6000Mhz RAM.  If this is for MSFS, then keep the CPU and go with 3080 GPU ... but be careful with eBay as there is a flood of GPUs that have been running 24/7 doing crypto mining.

EDIT: if you go with AMD, make sure you get good RAM ... performance difference can be as much as 19% just by RAM alone.  Intel seems to be less impacted by RAM speeds/latency.

Cheers, Rob.

Thank you for your Feedback Rob!

Honestly, I am not changing my mainboard nor my processor, however the PSU and GPU will be replaced. Especially eBay is crowded with suspicious persons who are trying to trick you to pay without any safety fees or directly send the money....however, the change between 10xx and 20xx is not that big, but especially the 3080-series are extremly powerful and in this case the 3080ti is a 3090 killer. After reviewing some benchmarks and tests it seems the 3080 will be the sweetspot.

Link to comment
Rob Ainscough

You'll probably not see a major difference in P3D with 3080/8700K combo.

However, you will probably see a significant difference in MSFS with 3080/8700K combo.

Just keep in mind the 8700K is a 5 year old CPU so don't expect miracles from it ;) 

Cheers, Rob.

Link to comment
Hendrik Herryjanto
On 10/1/2022 at 12:01 PM, Rob Ainscough said:

uc?export=view&id=1k1m04BID_Ll_YDiNzEU2C

Arrived

Hi Rob,

Can you share with us your result with this new gen AMD. Is it worth the upgrade?

  • Like 1
Link to comment
Rob Ainscough

I will, but I've had some issues with Win 11 22H2 that need to get sorted out first ... this Microsoft OS update has been a disaster for me.  Stay tuned.

  • Like 1
Link to comment
Markus Burkhard
On 10/7/2022 at 2:14 PM, Alexander Polcher said:

Will the Ti version anyhow be bottlenecked or slowed down by my i7 8700k and a 3080 would be sufficient, or would the Ti version bringt a noticeable performance boost?

Adding to what Rob has said, obviously your CPU/Mainboard will clearly bottleneck both GPUs, 3080 and Ti. The 3000 series GPUs are considerably more advanced than the CPU/RAM/SSD combo you have installed, that hardware is just not able to shovel data to those GPUs fast enough to unlock their full potential.
But you will of course still get a performance improvement, as Rob has pointed out well.

Having said that, I would recommend getting a 3080 Ti if you can find one at a good price, just to future-proof a bit better with those 2GB additional VRAM. Sure with your current other hardware a 3080 would be sufficient to boost your sim, but I reckon you will change your remaining hardware at SOME point in the not too distant future, so why not make your new GPU a bit more relevant for that time.

  • Like 1
Link to comment
Rob Ainscough

Managed to sort out my Win 11 22H2 issues, a few OS updates later and DDU and manual driver install.

Also managed to secure an ASUS 4090 OC edition this morning due to arrive Oct 17th ... I know I said I wouldn't go nVidia, but the performance numbers of the 4090 just too good to pass up even on the raster side.  Hopefully my 1600 WATT PSU can handle it ... eventually I'll get a water block for it from EK.

 

 

Asus4090OCEdition.jpg

Link to comment
Ray Proudfoot

Rob, any chance you could give some info on how well P3Dv5 runs with your new hardware?

I have zero interest in MSFS. If you can it would be greatly appreciated.

Link to comment
Wilhelm Zwirchmayr
3 hours ago, Rob Ainscough said:

Managed to sort out my Win 11 22H2 issues, a few OS updates later and DDU and manual driver install.

Also managed to secure an ASUS 4090 OC edition this morning due to arrive Oct 17th ... I know I said I wouldn't go nVidia, but the performance numbers of the 4090 just too good to pass up even on the raster side.  Hopefully my 1600 WATT PSU can handle it ... eventually I'll get a water block for it from EK.

 

 

Asus4090OCEdition.jpg

 

Yes, the 4090 is a beast

I looked at a few test reports and a well-known youtuber from the German area "Der8auer" - no spelling mistake

It is best to set the TDB to a fixed value or only 70% of what is specified by the manufacturer

You may have max. 5% less FPS but you need less power than a 3080 TI and you are still extremely fast

  • Like 2
Link to comment
Rob Ainscough

Absolutely, that's my plan.  I'll do both P3D and MSFS for testing.  I'll get to work on baseline (fully loaded sims) for both using my existing setup (5950X with 3090) this weekend.  If time permits I'll also do baseline without add-ons on multiple computers (9900K, AMD 5950X, AMD 7950X using various GPUs Titan X, Titan RTX, 3090, 4090).

I'm still waiting for EK CPU water block for the AMD 7950X, claim Oct 21st shipping ... that's currently the only hold up.  I can do the EK 4090 GPU water block at a later date.

Work has me swamped at 14 hr days recently (production issues and legacy apps issues), hopefully that will settle down soon and can get back to normal days.

On a side note, I had no serious buyers for my SimPit so figured I'll just keep:

uc?export=view&id=1p01sWcqQELtLm6GS80Tjh

or for those blue days ...

uc?export=view&id=1W9MIufQJCkw5L7hhuViYJ

I save purple for the wife when she operates as captain (oh wait, that's all the time).

Cheers, Rob.

  • Like 1
Link to comment
Ray Proudfoot

Rob, many thanks for including P3D in your testing. That’s one monster of a setup.

Link to comment
Rob Ainscough

P3D will be tested first since it's my primary platform, MSFS is still not where I want it to be and may never be ... I guess I'm just not a significant target to reach for MSFS project scope.

Cheers, Rob.

Link to comment
Rob Ainscough

Looks like the 4080 didn't do so well in this review set:

Heavy reliance on DLSS 3.0 support.

I'm still very interested in the AMD 7000 series GPU to see what they have to offer.

EDIT: 4090 shipped with tracking, ETA 17th, so it'll probably be the following weekend before I can start the build process then following weekend 29th before I can get some test numbers.

On a side note: the AM5 Asus Crosshair Extreme MB is by far the heaviest motherboard I've ever purchased.

Ray, I tried going back to standard clouds (Volumetric clouds off in P3D), kept EA On, loaded up some REX cloud structures and I've forgotten just how much more real the old sprite layers look, nice cirrus too ... their only really drawback is when you look directly down on them they spin.  In addition the old school sprite clouds light up when flying thru them with DL enabled. 

 

  • Like 1
Link to comment
Rob Ainscough

To add, not a fan of DLSS any version ... rather than upscale the image to higher resolution and then down sample (which is SSAA), DLSS will down scale resolution and fills in the "missing" pixels ... not really missing just makes an "educated" guess and that guess is what nVidia call AI.  It's more closely associated to a 4K TV that upscales 1080p broadcast image with primary difference being that it does not have a base 4K frame from which to reference for the "AI" processing.  AI is a stretch in definition.

From the video reviews I've seen of the 4080 12GB and 16GB are barely on par with 3090 and in some cases worse when DLSS is NOT used.  It's only when DLSS 3.0 is supported does the 4080 do much better than the older 3090.  

The 4090 does seem to perform better (without DLSS) but clearly not 2X better, more like 30-50% pending game (which is still a great improvement).  In MSFS the 4090 review numbers I saw were 26% higher FPS than 3090Ti (without DLSS).  When I get to the point of doing my own testing, I'll be sure to load P3D and MSFS and XP12 up with something more than a Cessna.

I'll do generic testing no add-on testing also as that provides a base for how well a sim performs "out the box", but I think most people are interested in FPS numbers are looking for improvements in a fully loaded add-on environment with more complex aircraft ... both GPU/CPU loading and finding the balance.

In P3D v5.x setting the various affinity options (more than just one now) combined with setting up add-on tasks to run on specific cores (via Lasso or other tools) is very beneficial and probably even more so with newer CPUs and DDR5 and PCIe 5.0 as more and more system wide bottlenecks are being addressed.  I haven't researched if MSFS or XP12 have the ability to define render cores (I think XP12 does), but I'll investigate during my testing and see if there is some optimization to be found in that area.

Cheers, Rob.

Link to comment
Ray Proudfoot
13 hours ago, Rob Ainscough said:

Ray, I tried going back to standard clouds (Volumetric clouds off in P3D), kept EA On, loaded up some REX cloud structures and I've forgotten just how much more real the old sprite layers look, nice cirrus too ... their only really drawback is when you look directly down on them they spin.  In addition the old school sprite clouds light up when flying thru them with DL enabled. 

I can’t have EA On. It darkens the VC of the Xtreme Prototypes Lear 25 too much and nothing I’ve tried lightens it sufficiently.

I’ll need to check my settings as I am happy with what I have for my daytime flights. I also found with EA On the sky never turned a deep blue and the bluish tinge of the ground from altitude looked like those old photos taken on negative film without a UV filter.

Link to comment
Ray Proudfoot

Hi Rob,

This site was recommended on AvSim as being a useful way of checking if your CPU and GPU were a good match for CPU or GPU intensive programs.

It suggests the i9-11900K and 3080Ti would be perfect for CPU intensive apps but for GPU intensive ones the 3080Ti would be the weak link.

https://pc-builds.com/bottleneck-calculator/result/10Y197/3840x2160/2/processor-intense-tasks

Thoughts?

Link to comment
Rob Ainscough
11 hours ago, Ray Proudfoot said:

I can’t have EA On. It darkens the VC of the Xtreme Prototypes Lear 25 too much and nothing I’ve tried lightens it sufficiently.

Did you try checking the Auto Exposure compensation?  That did the trick for me on those dark cockpits.

I like EA but it does tend to over haze/scatter all the time not allowing for days that can be very clear.  I've asked if LM would provide a simple p3d.config option where we can adjust the haze level but they said no.  EA also doesn't always interact well with non PRB terrain/surfaces, so legacy scenery that isn't PBR will look a little odd.  But you can get the deep blues, you just have to be at very high altitudes.

I'm back to the layered sprit clouds, 99% of the time they look excellent.

Link to comment
Ray Proudfoot
3 hours ago, Rob Ainscough said:

Did you try checking the Auto Exposure compensation?  That did the trick for me on those dark cockpits.

Yes, tried everything but the Lear25 VC is black so maybe it couldn't be improved. It's no great loss given I don't like the effects of EA and in any case it's quite hard on my old 1080Ti.

Incidentally, I've just bought Flightbeam KPDX. Spectacular scenery in that part of the world  - even in P3D. :)

Link to comment
Rob Ainscough

I think Orbx fixed Portland Cityscape also ... I know incredible Orbx actually spent a minute to fix something in P3D :) 

I asked if they were interested in having ALL their assets converted over to PBR materials but never got a response.  I would have done it for free and let them keep ownership and even sell the work, but still no response.  Maybe when AO comes to P3D, I'll ask again.

Yes, we have a lot of nice scenery, rivers, lakes, mountains, trees (so many trees) downtown Portland:

uc?export=view&id=1Dapfi4fE-5TBtIgbjyjLM

Lake Trillium

uc?export=view&id=1Nle_UHVvNGArw2FJ9zj_j

The beach

uc?export=view&id=1eO_ig4-qTsjoGJb2_9F4t

I do miss the UK, especially Scotland. 

On another side note, my wife's company have invited us on their private company business Jet for company Xmas party in Santa Barbara so I'm hoping to get some flight time (not flying, just observing) in the cockpit and video of course since there are less restrictions (not commercial).  Need to get exact details of bizJet type, 14 seater.

TGIF!

EDIT: Biz Jet is a Cessna 560 Citation registration N731TR,  housed at KUAO, our route will be KUAO to KSBA.

  • Like 6
Link to comment
Rob Ainscough

Most of the parts have arrived but still waiting on the EK water block for AM5 due to arrive Oct 26th.  The EK full front/back coverage GPU water block is due to arrive Nov 17th according to latest EK update.

The 4090 arrived today ... it's 14" long and 6.1 lbs.

nVidia4090-2.thumb.jpg.da9b96e2b4ac8007f6dfbcf5b5317ebe.jpg

nVidia4090-1.thumb.jpg.3ecaf2395d8ecb03a961b536eb996974.jpg

Has 3 X DisplayPort 1.4 and 2 X HDMI 2.1 ports

nVidia4090-3.thumb.jpg.2cda9ae34b35bf32e71009aab8748671.jpg

Asus Crosshair X670E Extreme motherboard (20 + 2 power stages) and 9.8 lbs

AsusCrosshairX670EExtreme2.thumb.jpg.8669dd474ec8fe39475c123564e4adca.jpg

G.Skill Neo EXPO DDR5 6000 CL30

GSkillDDR56000C30.thumb.jpg.08b532255e2dc700e849d8350e23da92.jpg

I've got my 6 test scenario's setup: 

P3DMSFSPerformanceTesting.thumb.jpg.2eb8dcf0d95cb7b016d93bc4f03af95c.jpg

I'll start with AMD 5950X / 3090 and then swap out 3090 for 4090 since I'm waiting on water blocks to arrive.  Again, this is NOT platform competition and I'll post graphic settings of both platforms.  I have no vested interest in anything on any platform and I'm a beta tester for both and equally harsh or complimentary to both platforms.  Sadly I don't think I'll have time to do XP12 in this round, but will once XP12 gets more stable and I have more add-on options to closely match my current two platforms.

If anyone is looking for a specific test on either platform, then let me know (no promises I'll do it).

Cheers, Rob.

 

  • Like 5
Link to comment
Rob Ainscough

I'll be using RTSS for logging at 100ms sample rate and MSI Afterburner for in-game overlay and assigned to run on their own cores so as to be minimal interference.  Video capture is performed by a separate PC via HDMI 2.1 4K splitter to 4K video capture card so as to introduce NO interference.

Would be great to test the Concord but I suspect I know one possible delay ;) (maybe). 

Cheers, Rob.

  • Like 3
Link to comment

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...