Search Results

Search found 29396 results on 1176 pages for 'multiple graphics devices'.

Page 79/1176 | < Previous Page | 75 76 77 78 79 80 81 82 83 84 85 86  | Next Page >

  • How would I go about implementing a globe-like "ballish" map?

    - by rFactor
    I am new to 3D development and I have this idea of having the game world like our globe is - a ball. So, there would be no corners in the map and the game is top-down RTS game. I would like the camera to go on and on and never stop even though you are moving the camera to the same direction all the time. I am not sure if this is even possible, but I would really like to build a globe-like map without borders. Can this be done, and how exactly? I am using XNA, C#, and DirectX. Any tutorials or links or help is greatly appreciated!

    Read the article

  • How do I adjust the resolution on an Intel gma x3100?

    - by Salinski
    I'm with fresh install of 12.04, having gma x3100 as a video card. The problem is I own 19" monitor with native resolution of 1280x1024 but can't force screen resolution on more then 1024x768 ~$ xrandr Screen 0: minimum 320 x 200, current 1024 x 768, maximum 4096 x 4096 VGA1 connected 1024x768+0+0 (normal left inverted right x axis y axis) 0mm x 0mm 1024x768 60.0* 800x600 60.3 56.2 848x480 60.0 640x480 59.9 I been digging up some info for the past 2 days, yet haven't found any solution. Even tried using gdm instead of lightdm. Any help is appreciated.

    Read the article

  • What Shading/Rendering techniques are being used in this image?

    - by Rhakiras
    My previous question wasn't clear enough. From a rendering point of view what kind of techniques are used in this image as I would like to apply a similar style (I'm using OpenGL if that matters): http://alexcpeterson.com/ My specific questions are: How is that sun glare made? How does the planet look "cartoon" like? How does the space around the planet look warped/misted? How does the water look that good? I'm a beginner so any information/keywords on each question would be helpful so I can go off and learn more. Thanks

    Read the article

  • How far can you get in iOS without learning PhotoShop or another graphic design program? [on hold]

    - by Aerovistae
    I'm in the process of learning iOS, and I'm coming from a web dev background where CSS controls 70-90% of the UI, and Python/C++ desktop dev where there are highly customizable UI toolkits for most things. I'm trying to figure out how people make good-looking apps without graphic design skills. You always hear about some 8 year old or 14 year old who made a successful app. So I assume that even if the required code was relatively basic, the app must have looked good if it was a success. But I find it really unlikely that these kids have advanced PhotoShop skills as well as having learned iOS programming at such a young age. Frankly, the same goes for most independent app developers....as they say, unicorns don't exist. So what's the deal? Can you make a good-looking, market quality app without those skills? What are the limitations?

    Read the article

  • Will 3D games run in a VirtualBox4-hosted Windows system under Ubuntu with Intel GMA X4500HD card?

    - by Ivan
    I've just got a laptop with Intel GMA X4500HD card (which is said to be DirectX10-compliant, while my previous laptop had Intel 82852/855GM and no chance to play modern 3D games at all). It would be nice if I could play some games (like HalfLife 2, StarCraft 2, Fallout 3) from time to time (while I am ok using lo-fi settings and having low performance) but as I strongly prefer Ubuntu environment and don't need Windows usually I'd like to install Windows inside VirtualBox instead of using a dedicated partition and dual-boot. So the question is if VirtualBox can use 3D acceleration of Intel GMA X4500HD if run on Ubuntu.

    Read the article

  • Ubuntu on low powered laptop

    - by zkent
    First off, I am new to Ubuntu. I come from a Windows background (DOS before that) and am loving it so far. I installed it on an older Dell laptop that I wanted to get another year out of. I set this machine up primarily as a LAMP development machine for a project I am working on. The machine is a Vostro 1500 and it has 4GB RAM (maxed) and an upgraded hard drive. I can't watch YouTube videos for long before it starts to overheat and start acting sporadic. I can live without YouTube but every so often the application switching (alt-tab) gets slower and begins not showing all applications and the Dash home quits displaying properly. I am sure I am asking a lot of this old machine. What I really want to know is: are there any settings in Ubuntu that allow me to lower the graphic effects (fade-ins, transparencies, fancy transitions, etc) that would be less taxing on the video card?

    Read the article

  • Creating a WARP device in managed DirectX

    - by arex
    I have a very old graphic card that only supports shader model 2, but I need shader model 3 or up for the app I am developing. I tried to use a reference device but it seems to run very slowly, then I found some samples in C++ that allows me to change to a WARP device and the performance is good. I am using C# and I don't know how to create such type of device. So the question is: how do I create a WARP device in C#? Thanks in advance.

    Read the article

  • Why is text on my TV monitor all grainy?

    - by searchfgold6789
    I have successfully installed the Catalyst drivers and hooked up my 1080p TV to the Radeon HD 5750's HDMI port and now it is working. There only seems to be one issue - most images appear OK, at least to my eyes, but text is grainy and difficult to read (not sure how well it can be seen in the picture): It works fine with a smaller screen, and hopefully you can see it with the new pictures which are only slightly better, but images are rendered a little funny too. Changing antialiasing does nothing and changing the refresh rate only makes a slight change if any. The small monitor I have works fine. Smaller resolutions such as 1024x768 look OK but are the wrong aspect ratio. I have tried renaming the input mode on the TV to "PC", and that helped... a little bit. Also, I tried a different PC with the same result. Messing around extensively with the ATI Catalyst Control Center also did not help. I am wondering if there is anything else I can try to increase the quality so it is more usable.

    Read the article

  • How do i add a start menu page to my java game?

    - by user2149407
    I have a rather cool space invaders game that my friend and I have been working on for a while, and we have decided it needs an opening page, with "Start" options, "Quit" options and so forth. I have looked at several methods online, but cant seem to get any of them to work! Does anybody have any ideas? P.S Using JFrame to draw the main frame Im just looking to do this within Java, so just a panel that appears at a state change (GAME, MENU). Id like it to contain a few buttons to start the game, and quit. Later, I will add achievements, but im after something really basic for now. But thanks for the suggestions!

    Read the article

  • Why doesn't my Nvidia driver support 1280*1024?

    - by Alvar
    EDIT: I think the problem is in the NVIDIA x server, that doesn't support an RGB screen from 1990. The screen was the best of them all when it was released. This a strange thing! In windows the EDID must work because 1280*1024 is used as max resolution, but in ubuntu those stuff aren't being received, so how do I tell ubuntu/nivida to find those EDID stuff? Hi, I have an external monitor which is an 20" 1280*1024. In windows 7 it works fine with that resolution but in ubuntu it can't. The screens model name is IntelliColor Display/20 and the name of the company is "radius".recived Well, twin view is working but the only maximum resolution is 1024*768. I'm using nividas x server program. If I force it to use 1280*1024 it all crashes and I have to force a reboot. alvar@alvars-laptop:~$ disper -l display DFP-0: HSD121PHW1 resolutions: 320x175, 320x200, 360x200, 320x240, 400x300, 416x312, 512x384, 640x350, 576x432, 640x400, 680x384, 720x400, 640x480, 720x450, 640x512, 700x525, 800x512, 840x525, 800x600, 960x540, 832x624, 1024x768, 1366x768 display CRT-0: CRT-0 resolutions: 320x240, 400x300, 512x384, 680x384, 640x480, 800x600, 1024x768, 1152x864, 1360x768

    Read the article

  • After 10.10 -> 11.04 upgrade, can only login via Classic (No Effects)

    - by Ryan P.
    Yesterday I upgraded from 10.10 to 11.04, everything seemed to go okay until immediately after login: the desktop goes into a "corrupted" looking state (similar to having too high resolution set). I can see some kind of movement by moving the mouse around/right clicking, and can enter text terminals via ctrl + alt + f1 It does this in both plain "Ubuntu" and "Ubuntu Classic", and only seems to login/startup properly with Ubuntu Classic (No Effects). I have checked my video card (Radeon X600) and run the unity support test which passes with all "yes" results (Unity supported: yes): /usr/lib/nux/unity_support_test -p I have tried re-installing my Ubuntu desktop: rm -rf .gnome .gnome2 .gconf .gconfd .metacity sudo apt-get remove ubuntu-desktop sudo apt-get install ubuntu-desktop With no success. I can workaround for now with Classic (No Effects), but I'd really like to find the root problem. Any suggestions on what else to try would be appreciated!

    Read the article

  • Help my graphists sharing their work

    - by Andy M
    As a developer I'm used to Subversion for source control and I think it's great for sharing source code between developers. Now thinking about my graphists and game designers, they need to have a slightly different approach I think. They need to share binary files They need to be able to have a thumbnail and preview of their work I don't want to include their binaries into my game repository (would be much too heavy for developer when updating) I've seen that some graphists uses personally created website to share their work but I was wondering if some "standard" application existed in order to provide my graphists a cool way of working together. Is there a common way of dealing with this? Is the way I want to do (only final sprites on my game repo) correct? How do you guys do this as game developers?

    Read the article

  • How can I efficiently create/store/implement animations as I add to my game?

    - by nickbadal
    My game's characters are made up of different parts (head/body/legs/etc), and whatever items they have equipped. As I'm creating the animation system for my game, I want to try to anticipate a large number of combinations for different pieces for each character. Originally, I had planned on having a frame-by frame animation for each piece for each animation, and then layer them to combine them into a character, but this seems like it would be a lot of work for my artist, and that the memory/disk size would start to add up as well since we would need a sprite for every frame, of every customization of every piece, in every animation, for every character. What efficient ways are there to create/implement these animations as we add more and more configurations to our game?

    Read the article

  • Pre-rendered fire. Where to find? [on hold]

    - by Vladivarius
    I'm studying game programming. I haven't yet implemented generated fire rendering in my ,,engine'' so I'm looking for some pre-rendered fire textures for early demo-scenes, but they seems strangely difficult to find. I'm currently using some that I ripped from DMC but I want to try out different ones. Does anyone know where to find these? Software that could generate them would also be ok. Thanks :)

    Read the article

  • Fog with Blend in OpenGL

    - by MhdAljobory
    I want to add fog in my scene which contain transparent textures made by Blend , when i enable the fog the transparent textures appear white From a distance but when i disable it the textures appear well. What is the solution to the problem of whiteness? Fog Code: GLfloat fogColor[4]= {0.5f, 0.5f, 0.5f, 1.0f}; glClearColor(0.5f,0.5f,0.5f,1.0f); glFogi(GL_FOG_MODE, GL_LINEAR); glFogfv(GL_FOG_COLOR, fogColor); glFogf(GL_FOG_DENSITY, 0.35f); glHint(GL_FOG_HINT, GL_DONT_CARE); glFogf(GL_FOG_START, 1.0f); glFogf(GL_FOG_END, 1000.0f); glEnable(GL_FOG); Screenshot

    Read the article

  • How to disable VGA

    - by Bitmap
    If i run lspci | grep VGA I get below output which tells me below VGA cards are present on my computer. 01:00.0 VGA compatible controller: nVidia Corporation GT216 [GeForce GT 220] (rev a2) 08:02.0 VGA compatible controller: ATI Technologies Inc ES1000 (rev 02) The ES1000 is an onboard card which came with my machine. Do anyone know how to disable this VGA on my machine. The reason for this request is because if I run xrandr I get the output as shown below: xrandr: Failed to get size of gamma for output default Screen 0: minimum 320 x 240, current 1024 x 768, maximum 1024 x 768 default connected 1024x768+0+0 0mm x 0mm 1024x768 50.0* 800x600 51.0 52.0 53.0 680x384 54.0 55.0 640x480 56.0 512x384 57.0 400x300 58.0 320x240 59.0 Which means I am not able to configure nVidia to accept smaller resolution. Thank you.

    Read the article

  • Trouble with FGLRX in quantal x86_64

    - by Dii Leoo
    it's the third time i have to reinstall quantal because the fglrx is making me sick. I have an ATI Mobility Radeon HD 4250 (the code ir RS880).When i install Ati Driver, i don't have any problems, but when i reboot, the resolution is minimal and the launcher isn't visible, i can only use the things that are in desktop and use the terminal using CTRL+ALT+T. i've tried to make specific distribution page, downloading the legacy driver from AMD, and another thing: when i had 12.04 jockey told me that i could install some drivers, i had trouble, but that doesn't matter. now, the option to install drivers doesn't tell me about any drivers, i mean, its blank. At the moment i'm using the free drivers (Gallium 0.4 on AMD RS880M. is there any alternative ? or i'll have to live with the free drivers ? Thanks a lot ! I KNOW I HAVE TO INSTALL LEGACY DRIVERS, I'VE INSTALLED THEM, THEY DON'T WORK.

    Read the article

  • Running crysis 2 on ubuntu gives me an error message

    - by ShajD
    so i recently installed ubuntu alongside windows 7 and i installed crysis 2 with wine. crysis 2 works fine when i run windows, however when i run it using wine in ubuntu cryengine gives me a message saying, "Unsupported video card detected! Continuing to run migth lead to unexpected results or crashes....." i've got two video cards ones an intel and the other's a nvidia. i typed lspci into the terminal and my nvidia card was listed under video controller as well

    Read the article

  • Blank Screen on 12.04

    - by stupidwhiteguy
    Can hit Ctrl-Alt-F1 and login but no fixes work. Tried nomodeset and everything else appears to have upgraded but can't see anything. I really need this computer to work. It is an old Dell with 1.6 processor with old school monitor. I don't know why people are giving my question negative votes. If you know how to fix this, then please help me, as I have read and tried all the solutions that I've found. Just for the record reformatting is not fixing it, and I can't get a boot CD to load either. If you know how to use the blank screen fix tutorial, please explain cos I'm lost. Most people don't understand this stuff and an easy "how-to" fix would be nice.

    Read the article

  • No screens found error with glasen/intel-driver

    - by pgcudahy
    A lot of people seem to have had success in getting hardware acceleration for intel 82852/855GM chipsets with the ppa:glasen/intel-driver. I've tried it on my Motion Computing M1400 but get a "no screens found" error. I've found one person out there with a similar problem who seemed to fix it, but his solution is in German and seems to involve recompiling the kernel (it's at the bottom of the comments). Anyone able to see how to fix this without such drastic measures?

    Read the article

  • Java single Array best choice for accessing pixels for manipulation?

    - by Petrol
    I am just watching this tutorial https://www.youtube.com/watch?v=HwUnMy_pR6A and the guy (who seems to be pretty competent) is using a single array to store and access the pixels of his to-be-rendered image. I was wondering if this really is the best way to do this. The alternative of Multi-Array does have one pointer more, but Arrays do have an O(1) for accessing each index and calculating the index in a single array seems to take one addition and one multiplication operation per pixel. And if Multi-Arrays really are bad, can't you use something with Hashing to avoid those addition and multiplication operations? EDIT: here is his code... public class Screen { private int width, height; public int[] pixels; public Screen(int width, int height) { this.width = width; this.height = height; // creating array the size of one index/int for every pixel // single array has better performance than multi-array pixels = new int[width * height]; } public void render() { for (int y = 0; y < height; y++) { for (int x = 0; x < width; x++) { pixels[x + y * width] = 0xff00ff; } } } }

    Read the article

  • Rendering trillions of "atoms" instead of polygons?

    - by Baring
    I just saw a video about what the publishers call the "next major step after the invention of 3D". According to the person speaking in it, they use a huge amount of atoms grouped into clouds instead of polygons, to reach a level of unlimited detail. They tried their best to make the video understandable for persons with no knowledge of any rendering techniques, and therefore or for other purposes left out all details of how their engine works. The level of detail in their video does look quite impressive to me. How is it possible to render scenes using custom atoms instead of polygons on current hardware? (Speed, memory-wise) If this is real, why has nobody else even thought about it so far? I'm, as an OpenGL developer, really baffled by this and would really like to hear what experts have to say. Therefore I also don't want this to look like a cheap advert and will include the link to the video only if requested, in the comments section.

    Read the article

  • Ask temperature on ubuntu 11.10 too hot

    - by stacheldraht
    I wonder why when I am using Ubuntu, my laptop temperature increases. It's around 62 degrees or less. I am already using Jupiter and have set it to power saving mode but if I use maximum performance the temperature can reach 70 degrees or more. Its too hot if I compare it when I am running windows where the temperature is quite normal at 56 degrees. How can I solve that?, sorry for my bad English.

    Read the article

< Previous Page | 75 76 77 78 79 80 81 82 83 84 85 86  | Next Page >