I'm not convinced this will help Xonotic. As the dev there states, AA is CPU bound, not GPU bound. Xonotic tends to be more GPU bound already so making this change may actually hit performance.
I'm at least a reasonably tolerable person to be around - Narcopic
@edh: If you would have a look at the benchmark results then you´d quickly find out that the GPU isn´t that much of an interest. The CPU still seems to be the key part of your setup. Having an intel i7 and an old geforce 9800 (or similar) gives you extremely high FPS whereas using an AMD dual core (e.g. 2x2GHz) and a GTX560 isn´t that ideal.
(09-24-2012, 08:22 AM)Maddin Wrote: @edh: If you would have a look at the benchmark results then you´d quickly find out that the GPU isn´t that much of an interest. The CPU still seems to be the key part of your setup. Having an intel i7 and an old geforce 9800 (or similar) gives you extremely high FPS whereas using an AMD dual core (e.g. 2x2GHz) and a GTX560 isn´t that ideal.
I think we have some different targets we are talking about here. If someone is getting 500fps on low details with the latest hardware, I'm not really bothered. This would be CPU limited on most systems. On the higher detail levels it is mostly GPU bound, otherwise you'd be getting the same 500fps all of the way up to Ultimate. When looking at older hardware you also see more of a GPU limit. Even comparing a Geforce FX5600 vs an FX5950 Ultra in a 2x1GHz Pentium 3 where each graphics card is way beyond the capability of the CPU, they each give drastically different performance which shows it to be GPU not CPU bound in such an environment. Similarly with a 6600GT in an Athlon XP 2400+ which is massively over powered on the GPU vs CPU there are very different framerates across the detail levels. This would not be the case if it was CPU bound. Making the game playable on older hardware is more of interest to me than someone getting 500fps on the newest hardware instead of 480fps. In addition there is a level of disparity I have reported on before between Windows and Linux which may show a level of GPU vs CPU bias: http://forums.xonotic.org/showthread.php...2#pid35982
I'm at least a reasonably tolerable person to be around - Narcopic
09-24-2012, 01:09 PM (This post was last modified: 09-24-2012, 01:10 PM by asyyy.)
No proof but at least evidence:
Old system:
AMD Phenom II x3 @ 2.9ghz
ATI HD5750
Tried overlocking my gpu, gained 0% fps (average around 150). Then upgraded my CPU to AMD Phenom II x4 @ 3.4ghz -> 25% fps increase. On high quality configs the GPU might be the bottleneck, but on normal-low configs the CPU certainly has a huge impact. Maybe some systems are able to run Xonotic at 500fps, but I know quite a lot people who have problems on getting stable 125fps on a low-ish config.
No AMD APUs here? Not to be underestimated, these are strong performers... at least the hig-end ones ( A8-5600K , A10-5800K )... so here are my results:
Platform:
=======
Processor : AMD APU A8-5600K ( 3.6Ghz , 4 x "Piledriver" integer cores, 2 x 256-bit FMAC units, 4 MB L2 Cache )
GPU : Radeon 7650D integrated on APU ( 256 shader units )
RAM: 8GByte HyperX 1600 ( at 667 Mhz ) memory, dual-channel.
Operating system info:
=================
OS : Slackware Linux 14
Arch : x86_64
Cores : 4
Memory: 7608580 ( 512 MB reserved for GPU )
GL information:
===========
Vendor: GL_VENDOR: ATI Technologies Inc.
Card: GL_RENDERER: AMD Radeon HD 7560D
Driver: GL_VERSION: 4.2.11931 Compatibility Profile Context
Extra info:
========
Re-compiled binaries with GCC 4.7 with flags: -march=bdver2 -msse4.2
Conclusions:
==========
Pretty nifty results , playing at 1280x768 on this is a blast! and the Keybench benchmark never drops of 60 FPS on Ultra settings ( Minus Decals, Minus Reflections, Particle Density = 0.7, but All shadowscheckboxes ON )
Considering harsh reviews coming from internet...AMD APUs ( higher end ones ) are a viable platform for gaming, unbeateable on price
ferndan, could you please include the log files too? And, since you recompiled the engine, tell us where you got the sources (git? if so, what revision?).
(12-18-2012, 10:39 AM)ferndan Wrote: Re-compiled binaries with GCC 4.7 with flags: -march=bdver2 -msse4.2
Hey ferndan, did you check the influence of these flags on your game performance? (Benchmark with a recompiled version, but without tuning for a special architecture.) I was always wondering how much impact these flags have in "everyday" applications
4. Add the flags you need, save and next execute :
make
5. It will generate THREE binaries:
nexuiz-dedicated
nexuiz-glx
nexuiz-sdl
6. Copy them on the Xonotic root folder... but rename them to the corresponding "xonotic" naming , depending on architecture ( backup the original binaries first ):
xonotic-linux32-dedicated or xonotic-linux64-dedicated
xonotic-linux32-glx or xonotic-linu64-glx
xonotic-linux32-sdl or xonotic-linux64-sdl
Hope that helps... now I will run with the original stock binaries, I'll let you know soon
01-21-2013, 04:57 PM (This post was last modified: 01-21-2013, 05:01 PM by edh.)
I've tested on a few other systems from the last decade or so.
1. Sempron 2800+ @ stock 2.0GHz
1Gb RAM
ATI Radeon 9550
GL_VERSION: 2.1.8545 Release
Windows XP Pro 32-bit
Results: 67/63/52/24/19/-/-
The Windows driver appears to have some graphical errors with warpzones which make things a bit wierd. It also doesn't support offset mapping so ultra nd ultimate were not tested. Can't recommend this setup even if it is a playable speed because of the graphical issues.
2. Athlon XP 2000+ Palomino @ stock 1.666GHz
768Mb RAM
Geforce FX5900 128Mb BIOS modded to FX5950 Ultra
GL_VERSION: 2.0.3
Windows XP Pro 32-bit
Results: 50/48/41/21/15/-/-
This card does not support offset mapping so again ultra and ultimate were not tested.
3. Athlon XP 2000+ Palomino @ stock 1.666GHz
768Mb RAM
Geforce FX5900 128Mb BIOS modded to FX5950 Ultra
NVIDIA driver
GL_VERSION: 2.1.2 NVIDIA 173.14.36
Arch Linux x86
Results: 52/48/41/20/15/-/-
As before but under Linux. Performance is about the same, still no offset mapping.
4. Athlon XP 2000+ Palomino @ stock 1.666GHz
768Mb RAM
Geforce FX5900 128Mb BIOS modded to FX5950 Ultra
Nouveau, Gallium 0.4
GL_VERSION: 1.5 Mesa 9.0.1
Arch Linux x86
Results: 49 on OMG, then graphical errors after this.
Not bad speed compared to the propreitary driver on OMG but on Low things all fell apart. Can't recommend Nouveau for Geforce FX still.
5. Athlon XP 2000+ Palomino @ stock 1.666GHz
768Mb RAM
Geforce4 MX440
GL_VERSION: 1.5.7
Windows XP Pro 32-bit
Results: 55/35/25/-/-/-/-
Surprising performance perhaps as it actually appears faster than a Geforce FX5950 Ultra on OMG! However this is skewed by it not supporting OpenGL 2.0 so what it is rendering is not strictly the same. Can't recommend such old hardware.
6. Athlon XP 2400+ Thoroughbred @ 2.025GHz
1Gb RAM
Radeon 9550 (250/400), BIOS modded to 9600 (325/400) then overclocked to 405/472MHz, it would give 536/530 but I backed it off for longevity!
GL_RENDERER: Gallium 0.4 on ATI RV350
GL_VERSION: 2.1 Mesa 9.0.1
Arch Linux x86
Results: 56/45/38/21/16/-/-
The free driver works rather well. No graphical issues seem with the ATI Windows driver but still offset mapping is not supported.
Uploaded all of these results in single ZIP file.
I'm at least a reasonably tolerable person to be around - Narcopic
System Name: Gateway LT4004u Netbook
CPU: Intel Atom N2600 @ 1.6 GHz
2 cores, 4 threads
2 GB of RAM, running at DDR3/800 (underclocked by CPU)
Windows 7 Starter 32-bit
Intel GMA 3600
8.14.8.1065 driver (OpenGL 3.0 support)
Benchmark ran at 1024x768 (I have registry hack enabled for downscaling display)
I ran the benchmark before the registry hack at 640x480 and was able to complete all 4 runs on high settings at 14 fps (Ultra didn't do so well). At 1024x768, I got through 2 runs on high before Windows restarted.
(02-27-2013, 02:30 PM)qorl Wrote: Many of u here are posting results without resolution settings!
I think you're a little confused here on the Big Benchmark. What you have run is the big keybench with your own settings.
The Big Benchmark runs the big keybench with automated multiple runs on multiple settings, some of which are specific to The Big Benchmark and which won't apply if you run the big keybench manually. All of these tests are done at 1024x768 as that is a resolution which almost any modern display can support. Therefore no one is listing the resolution unless they are running something different for technical reasons, like on a netbook with a lower resolution.
Your results can not be counted as they are incomparable.
Please reread the first post in this thread, then rerun The Big Benchmark as described and post your results.
I'm at least a reasonably tolerable person to be around - Narcopic
So Geforce4 MX440 is better than a Radeon 9000 but still not playable. Really can't recommend hardware of this age as it does not support OpenGL 2.0 and hence maps that use warpzones are somewhat difficult.
I'm at least a reasonably tolerable person to be around - Narcopic
03-14-2013, 12:32 PM (This post was last modified: 04-09-2013, 01:26 PM by edh.)
(03-14-2013, 09:08 AM)qorl Wrote: Swift shader,
Will this software work on xonotic?
Why not try it? Better that way rather than getting someone else to view a Youtube video then hazard a guess. I don't tend to click on links to videos without any explanation - a one paragraph description uses much less bandwidth.
All it is is a software renderer. These have existed for a long time.
It is only better if there is no dedicated GPU and you only have really terrible onboard graphics but a massively fast CPU.
Their claim of 620 on 3DMark06 with an i7 really shows that you're not going to get much performance. From memory I used to get a few thousand in 3DMark06 with a 6600GT which is a reasonably playable card for Xonotic on normal details. I would therefore guess that SwiftShader will not get anything playable with pretty much any system.
This is worse performance wise that the integrated software renderer in Xonotic. It is already of some use for those without a dedicated GPU but it does require SSE2. You can try this if you want from the Xonotic console:
Code:
vid_soft 1
vid_restart
Sorry to break the news but there are no magic beans to improve performance like this. Their '100x faster' boast is only from something very, very terrible.
I'm at least a reasonably tolerable person to be around - Narcopic
I'm just some regular dude that plays xonotic;-)
My belief in all this graphic stuff that works good or bad is here just because of money, so I think that the answer lies in software.
I remember times when i was plying quake 1 in 1999 and was working fine on 8mb graphic, today all of this stuff is abolished by some
shader models. I'm not a professional in programming or whatever but i could see a scam right away.
Thx edh;-)
I'm QorL not a normal dude;-)
03-14-2013, 02:18 PM (This post was last modified: 03-14-2013, 02:25 PM by rafallus.)
@up
Not even close to being a scam, you are VASTLY underestimating, what GPUs now need to deal with. Sheer amount of data to process has gone up by several orders from magnitude (and if ray tracing becomes a practical thing, it will grow even further). You get better image in quality in return (more vertices means more detailed models, bigger textures too, AA/AF to deal with artifacts of rendering) Also, Quake 1 in 1999 wasn't exactly news, not even Quake 2, Q3A was released by the end of that year. And good luck playing Q3A on the card you're talking about.
Shader models or new OGL versions are not abolishing anything, in fact they are adding ways to achieve new effects not available at the time before. And as noted before, you need more computing power to deal with it.
03-14-2013, 03:09 PM (This post was last modified: 03-14-2013, 03:11 PM by edh.)
(03-14-2013, 01:18 PM)qorl Wrote: My belief in all this graphic stuff that works good or bad is here just because of money, so I think that the answer lies in software.
Xonotic uses OpenGL which runs in hardware on the graphics card. No magic software will improve that situation if you have a decent graphics card and driver that provides hardware acceleration.
To demonstrate, here are two comparisons:
1. Take a Core i7 920 with onboard graphics that provide no hardware acceleration (as they did in their testimonial) and install this software and you will experience massive improvements as they quote for 3DMark06. However, it still only gets 620 which is on a par with a Geforce FX perhaps.
2. Take the same system with the highest end gaming graphics card you can find and try installing this software. How much difference will it make? None at all. This is because the graphics card already provides hardware acceleration which will be much, much faster than any software path unless you really do have a supercomputer.
Software emulation can not match hardware acceleration.
(03-14-2013, 01:18 PM)qorl Wrote: I remember times when i was plying quake 1 in 1999 and was working fine on 8mb graphic
The graphics card is irrelevant to the original Quake so long as you have 2Mb video RAM and VESA 2.0 compliance. Both DOS Quake and WinQuake run entirely in software. The only difference that might be seen are when running a VESA 1.0 card as opposed to VESA 2.0 or running on a card with less than 2Mb RAM. VESA 2.0 cards provide faster 2D rendering (all the card does in Quake, the 3D is done in the engine) and you need greater than 1Mb to render the top resolution: 1280x1024.
I've been there before. Swapping a Geforce2 MX into an old Pentium and getting no performance improvement at all versus the Matrox that was in there before. Why? WinQuake only used software rendering. GLQuake and it's derivatives will make use of hardware so that it where the improvement was to be had and it looked better. Therefore rather than running 512x384 at 21fps I was able to run 1024x768 at 32fps.
(03-14-2013, 01:18 PM)qorl Wrote: i could see a scam right away.
It's not that it's a scam. It's a wonderful piece of software if you don't have a graphics card/driver which provides hardware acceleration. This may be an acceptable solution for light gaming uses like the Sims, compositing window managers and similar but for a graphically complex game like Xonotic where a dedicated graphics card is an accepted recommendation, this will achieve nothing, especially when we already have a software rendering path.
A bottom of the range graphics card is not a major expenditure and will outperform this substantially.
Besides, the software costs money and is closed source, hence even if it was to improve performance (which it won't for 99.99999% of people who could actually run Xonotic anyway), it can not be integrated into Xonotic.
I'm at least a reasonably tolerable person to be around - Narcopic