Search Results

Search found 18029 results on 722 pages for 'stripe size'.

Page 447/722 | < Previous Page | 443 444 445 446 447 448 449 450 451 452 453 454  | Next Page >

  • How can I improve the "smoothness" of a 2D side-scrolling iPhone game?

    - by MrDatabase
    I'm working on a relatively simple 2D side-scrolling iPhone game. The controls are tilt-based. I use OpenGL ES 1.1 for the graphics. The game state is updated at a rate of 30 Hz... And the drawing is updated at a rate of 30 fps (via NSTimer). The smoothness of the drawing is ok... But not quite as smooth as a game like iFighter. What can I do to improve the smoothness of the game? Here are the potential issues I've briefly considered: I'm varying the opacity of up to 15 "small" (20x20 pixels) textures at a time... Apparently varying the opacity in this manner can degrade drawing performance I'm rendering at only 30 fps (via NSTimer)... Perhaps 2D games like iFighter are rendered at a higher frame rate? Perhaps the game state could be updated at a faster rate? Note the acceleration vales are updated at 100 Hz... So I could potentially update part of the game state at 100 hz All of my textures are PNG24... Perhaps PNG8 would help (due to smaller size etc)

    Read the article

  • Rendering a DOM across multiple displays

    - by meetamit
    I'm building a data-driven animation with HTML and javascript to run in a web browser. I would like to display it tiled across three 1080p monitors. This essentially yields a viewport that's 5760px wide and 1080px tall. Pretty large. Does anyone have experience setting up something like this? I have many questions below, but any tip would be appreciated: Is it reasonable to expect a DOM to render into such a large viewport size at close to 60fps? I might choose to use canvas, instead of SVG or HTML, but that would yield a giant canvas. Can a canvas with such high resolution be performant? Of course everything depends on the complexity of the graphics I want to render, but I'm looking to remove that factor from this question, so assume I'm asking about a canvas animation that can run at 60fps at 1920x1080 resolution. Would it run roughly as fast at 3 times the width? Would three.js and WebGL be a more proper approach at that resolution? How do you actually cause Chrome or FF to span 3 monitors at full screen? Do I need a 3rd party solution of any kind? Thanks!

    Read the article

  • Problem in multi booting Ubuntu 12.04 with existing Windows XP, 7 and 8 in 500GB HDD with 5 Partitions

    - by Dhruva
    Here's my case. I have 500GB HDD with 5 Partitions with XP, Windows7 and Windows8 RP in the first three. As per one of the instruction I've seen in this forum, I did shrink my 4th Partition to create a 30GB unallocated free space to install Ubuntu 12.04. But, when next I'm trying to boot the Ubuntu CD and choosing "Something Else", its only recognizing my 500GB HDD in whole as "/sda" and not reading the free 30GB space separately to install Ubuntu in it as suggested in the instruction mentioned in this forum. I've also tried to install in from within Windows7, by mounting the Ubuntu ISO file and using the .exe file and instruction thereupon (choosing free drive, user name, installation size, etc.), but that also failed after the PC restarted to continue the installation, showing as error for file extension, partition something error. One thing to be noted that the PC I'm trying to install Ubuntu in it is my Home PC and doesn't have any internet connection. Hence, no updates or otherwise online help. What shall I do?? Kindly suggest. Sorry if I made some grammatical mistakes as English is not my first language. Thanks in advance.

    Read the article

  • Tail-recursive implementation of take-while

    - by Giorgio
    I am trying to write a tail-recursive implementation of the function take-while in Scheme (but this exercise can be done in another language as well). My first attempt was (define (take-while p xs) (if (or (null? xs) (not (p (car xs)))) '() (cons (car xs) (take-while p (cdr xs))))) which works correctly but is not tail-recursive. My next attempt was (define (take-while-tr p xs) (let loop ((acc '()) (ys xs)) (if (or (null? ys) (not (p (car ys)))) (reverse acc) (loop (cons (car ys) acc) (cdr ys))))) which is tail recursive but needs a call to reverse as a last step in order to return the result list in the proper order. I cannot come up with a solution that is tail-recursive, does not use reverse, only uses lists as data structure (using a functional data structure like a Haskell's sequence which allows to append elements is not an option), has complexity linear in the size of the prefix, or at least does not have quadratic complexity (thanks to delnan for pointing this out). Is there an alternative solution satisfying all the properties above? My intuition tells me that it is impossible to accumulate the prefix of a list in a tail-recursive fashion while maintaining the original order between the elements (i.e. without the need of using reverse to adjust the result) but I am not able to prove this. Note The solution using reverse satisfies conditions 1, 3, 4.

    Read the article

  • Few questions on giga tweaker

    - by user23950
    I better consult first the people here before I do anything unnecessary using this app called giga tweaker. I don't really understand what this increase the performance of your CPU thing. It is under Customization-Memory Management-ram & disk cache of giga tweaker. What will happen if I change the level cache size of l2 cache into the highest possible value which is 8Mb. What are the negative effects of doing it? The file system caching memory, still under Customization-Memory Management-ram & disk cache. What effects will it have on my system which has 2Gb of Ram and 2.50 Ghz of Dual Core CPU. Please enlighten me.

    Read the article

  • Virtual Box - not filling entire screen

    - by jdavis
    I am new to VirtualBox and am trying to set up an instance of Windows 7 64. I have the virtual machine instance running with Windows 7 now installed, but it only fills up a small portion of my screen. Even when I go full screen, the window stays the same size and the rest of the screen is filled with gray space. I have installed VirtualBox Guest Additions, which allowed me to go from a resolution of 800x600 to 1024x768, but this still isn't satisfactory as my laptop display is 1600x900. Any help on this would be most appreciated. Thanks.

    Read the article

  • Why can't gif images copy at a reasonable speed on this dell laptop with XP?

    - by alt234
    I've got this somewhat old Dell Latitude D810. Strangest thing... If I try to copy anything that has gif files in it the gif files take forever. Like a few minutes per gif regardless of size. Everything else copies fine. I notice this when copying files off our network, copying off multiple external drives, and even when files are copying during an installation process. I'm on Windows XP Pro service pack 3. I've never seen anything like this before. Anyone else?

    Read the article

  • Does lshw list the "factory" speed of a memory module or the effective speed and how to find the former?

    - by Panayiotis Karabassis
    I hope I phrased this correctly. lshw gives: description: DIMM Synchronous 400 MHz (2.5 ns) product: M378B5773CH0-CH9 vendor: Samsung physical id: 0 slot: DIMM0 size: 2GiB width: 64 bits clock: 400MHz (2.5ns) And indeed the memory speed is set is set to 800MHz in the BIOS, which I think makes sense since it is a double rate. On the other hand, Googling strongly suggests that to this product number corresponds the PC3-10600 type, which is 1333MHz, not 800MHz. And this seems to be confirmed in the BIOS, where if I select Auto for memory bus speed, 1333MHz is selected "based on SPD settings". However in the latter case, the computer does not boot, i.e. the kernel panics, complaining that something attempted to kill the Idle process. So, I am I am beginning to suspect that I have been given defective memory, the technician that installed saw this, and lowered the bus speed. Is this a possibility?

    Read the article

  • windows 2008r2 SNMP hrStorageSize reporting wrong values

    - by Giannis Nohj
    I'm trying to monitor disks on windows 2008 R2 system, but I'm getting wrong values returned from hrStorageSize and hrStorageUsed for one of the 2 disks the system has. Local disk C: is returning correct values, whereas D: disk, which is an SAN disk, is reporting 32G instead of 500G. Output of snmpwalk -v 2c -c public hostname-01 .1.3.6.1.2.1.25.2 : HOST-RESOURCES-MIB::hrStorageDescr.1 = STRING: C:\ Label: Serial Number 7ff87ca6 HOST-RESOURCES-MIB::hrStorageDescr.2 = STRING: D:\ Label:Data Serial Number dab5b0b9 ... HOST-RESOURCES-MIB::hrStorageAllocationUnits.1 = INTEGER: 4096 Bytes HOST-RESOURCES-MIB::hrStorageAllocationUnits.2 = INTEGER: 4096 Bytes ... HOST-RESOURCES-MIB::hrStorageSize.1 = INTEGER: 35834383 HOST-RESOURCES-MIB::hrStorageSize.2 = INTEGER: 7864320 ... HOST-RESOURCES-MIB::hrStorageUsed.1 = INTEGER: 32460169 HOST-RESOURCES-MIB::hrStorageUsed.2 = INTEGER: 3192302 I have calculated the first disk (C:) size as 136G, which is correct. On the other hand, D: is calculated as 30G, where it should be 500G. I also tried fsutil on the system and the output was correct. Anyone have any ideas?

    Read the article

  • Post raid5 setup reboot shows single hard drive failure on ubuntu 12.10?

    - by junkie
    I just set up raid 5 on linux using three HDDs as per a guide. It all went fine until when I rebooted I got the following text: http://i.stack.imgur.com/Zsfjk.jpg. Does this mean one of my HDDs has failed? How do I check if any of them are failing? I tried using smartctl and didn't see any issues. Or is it nothing to do with failure and something else altogether? I would like to get the raid 5 working again but I'm not sure where to go from here. I'm using ubuntu 12.10 and the three raid disks each have a gpt partition with a single full size partition of filesystem type ext4. Note I only got an error on reboot not while I was creating the raid array which went fine. Thanks.

    Read the article

  • Why is my disk full?

    - by Agmenor
    I installed Ubuntu 12.04 by doing a fresh install where there was previously Ubuntu 11.10. My computer warns me now that my disk is nearly full. After having run apt-get purge, run apt-get autoremove and emptied the Trash can, I still have this problem as shown by this screenshot of Gparted: The disk /dev/sda7 is indeed full. I ran the Disk Usage Analyzer (Baobab) and I am still not sure of what is happening: One of my hypothesis is that when installing Ubuntu 12.04, I didn't configure my disks well and the disk /dev/sda6 is not mounted well as /home. Is this the reason indeed? What should I do to verify this and then to get the things fixed? Here are a few additional details to answer the questions I received (thank you everybody): My home directory is not encrypted. The Backup utility (Déjà Dup) is not set for automatic backups. (I do it myself and manually.) After I mount /dev/sda6, the command df -h gives Filesystem Size Used Avail Use% Mounted on /dev/sda7 244G 221G 12G 96% / udev 3,9G 4,0K 3,9G 1% /dev tmpfs 1,6G 904K 1,6G 1% /run none 5,0M 0 5,0M 0% /run/lock none 3,9G 164K 3,9G 1% /run/shm /dev/sda6 653G 189G 433G 31% /media/8ec2fa69-039b-4c52-ab1b-034d785132a1 (sorry but formatting this into code does not work, for an unknown reason) Thanks to izx's post, I realized /dev/sda6 was not even mounted before. It contains all the documents I used to have when I was running Ubuntu 11.10.

    Read the article

  • Series On Embedded Development (Part 3) - Runtime Optionality

    - by Darryl Mocek
    What is runtime optionality? Runtime optionality means writing and packaging your code in such a way that all of the features are available at runtime, but aren't loaded and used if the feature isn't used. The code is separate, and you can even remove the code to save persistent storage if you know the feature will not be used. In native programming terms, it's splitting your application into separate shared libraries so you only have to load what you're using, which means it only impacts volatile memory when enabled at runtime. All the functionality is there, but if it's not used at runtime, it's not loaded. A good example of this in Java is JVMTI, Java's Virtual Machine Tool Interface. On smaller, embedded platforms, these libraries may not be there. If the libraries are not there, there's no effect on the runtime as long as you don't try to use the JVMTI features. There is a trade-off between size/performance and flexibility here. Putting code in separate libraries means loading that code will take longer and it will typically take up more persistent space. However, if the code is rarely used, you can save volatile memory by including it in a separate library. You can also use this method in Java by putting rarely-used code into one or more separate JAR's. Loading a JAR and parsing it takes CPU cycles and volatile memory. Putting all of your application's code into a single JAR means more processing for that JAR. Consider putting rarely-used code in a separate library/JAR.

    Read the article

  • Belarc Advisor (Store Passwords using Reversible Encryption)

    - by Steve
    Hi, I'm using Belarc Advisor to examine my PC. Part of BA is a security benchmark summary, which examines components of windows security and provides a benchmark rating. Two items are marked as Fail: - Store Passwords using Reversible Encryption - Password History Size I have opened the Local Security Settings tool from the Control Panel Administrative Tools, and ensured that the "Store passwords using reversible encryption" setting is enabled. Also, I've set the password history to a number. So I'm a bit miffed about the Fail marks. Any idea why the Fail marks appear? Any clues how I can Pass them? Thanks, Steve.

    Read the article

  • How do I partition my Hard drive to install Kubuntu?

    - by Xdflames
    I am a complete newbie to partitioning and I would like some guidance here. Can anyone explain what exactly I should be doing here? I am installing Kubuntu 12.04 on a currently Windows 7 laptop. My current Partitions say this: /dev/sda /dev/sda1 ntfs 104 MB 35 MB /dev/sda2 ntfs 319965 MB 87164 MB Basically, what I want is some guidance on what exactly to do here. All I want is a partition for my OS (Which will be Kubuntu 12.04) and a partition for all of my data. I also want to restart fresh with my hard drive, with only what I mentioned on here. I am installing Kubuntu from a flash drive (I set it up as a bootable device with Universal-USB-Installer-1.9.0.9), as I do not have any blank CDs/DVDs to burn to. What should I name my partitions? What should I set the type as? What size do they need to be? Edit: My hard drive is 320GB. Just looked it up in my BIOS. This computer will mainly be used for internet browsing and overall just messing around with the Linux OS.

    Read the article

  • How do I make an encrypted disk image on Debian?

    - by Blacklight Shining
    I'm basically looking for an equivalent to OS X's encrypted sparsebundles. The solution should have support for file ACLs and should not force me to specify a size in the beginning (the image should only take up as much space as it needs) or require root access to mount and unmount. Ideally, I should be able to set two different passwords (both for the same data), but that's not too important. (I do have root access to the machine and so can install packages and such, but I would rather not have to sudo just to mount an image.)

    Read the article

  • Should I format USB sticks and SD cards to FAT, FAT32, exFAT or NTFS? (Windows files, live Linux distors)

    - by superuser
    Does it depend on the media size which one to chose or on some other parameters? In Windows 7 FAT16 is the default. In pendrivelinux.com's Universal USB Installer FAT32. Which one to chose? How about NTFS for Windows use? How about exFAT? It is tne Microsoft designed filesystem for removable media. Is there a difference in USB sticks and SD cards in this regard? Edit: seeing developments in the other thread, should I still use something like exFAT if I don't want Recycle bins created on every single machine I plug my USB thumb drive in?

    Read the article

  • Moving tool bars in Acrobat X Pro

    - by user67275
    I'm using Acrobat X pro for Windows. I found that two lines of tool bar takes too much space. So I would like to combine them into one line, or move them to left or right side. I do know that I can hide those tool bars by Ctrl+H. But I want to use them without hiding. Since most documents I use are letter size and portrait type, tool bars on the left or right side would be much more efficient way of using space. Does anybody know how I can do this? (I think I could do it in Acrobat 9..but I can't in X version)

    Read the article

  • BUILD 2013 Session&ndash;Alive With Activity

    - by Tim Murphy
    Originally posted on: http://geekswithblogs.net/tmurphy/archive/2013/06/27/build-2013-sessionndashalive-with-activity.aspx Live tiles are what really add a ton of value to both Windows 8 and Windows Phone.  As a developer it is important that you leverage this capability in order to make your apps more informative and give your users a reason to keep opening the app to find out details hinted at by tile updates. In this session Kraig Brockschmidt cover a wide array of dos and don’ts for implementing live tiles.  I was actually worried whether I would get much out of this session when Kraig started it off with the fact that his background is in HTML5 based apps which I have little interest in, but the subject almost didn’t come up during his talk.  It focused on things like making sure you have all the right size graphics and implementing all of the tile event handlers.  The session went on to discuss the message format for push notification and implementing lock screen notification and badges. As with the other day 1 sessions it was like drinking from a fire hose, but it was good stuff.  Check it out when they post it on Channel 9. del.icio.us Tags: BUILD 2013,Live Tiles,Windows 8.1

    Read the article

  • nginx static file buffer

    - by Philip
    I have a nfs which several frontend-servers are connected to for making the files stored on the nfs available for http downloads. It looks like I have problems with the way apache is serving the files, there seems to be a very small buffer or no buffer at all which results in a lot disk seeks. I did some testing with loading the whole requested file into memory at once and serve it to the client from memory. With this technique I need less disk seeks for a download stream. Since I don't want to implement this by myself for production use I thought that I could maybe use nginx for that because the documentation says that it uses buffers for static file serving. Is it possible to increase the buffer size to a few mb, if so which config parameter do I have to change for this? Has anyone experience with large buffers for static file serving? Is there a better way to reduce disk seeks?

    Read the article

  • Handout export to word from PowerPoint are too big :(

    - by nickjohn
    EDITED i am using power point lectures. i want to mail merge speaker data into the respective lecture. now thats not possible with ppt as far i know, so i have to convert these lectures to handout by using power point option "publishMS word handouts" and use word mail merger. this is good since it will keep the comments/notes added in slides in handouts aswell. but these exported handouts in word remain actual slides and retain link to original ppt rather than simply get exported as images. so the file size gets verrry big 10mb ppt = 212mb doc=88mb docx Is there any option to convert handouts exported from power point to word as images? i simply cant save them as pngs from powerpoint since that will not include the comments data. Thanks

    Read the article

  • Install Windows7 on drive with Ubuntu 12.04 already on. Is my plan good?

    - by John F
    I have Ubuntu 12.04 working fine, but need W7 occasionally. I just wanted to check that my plan for installing would work? Any help appreciated. Current partitions are: Partition....@ File System @ Mount Point @ Size.....@ Used.....@ Flags /dev/sda1....@ ext4........@ /ext4a......@ 37 GiB...@ 776 MiB..@ boot /dev/sda2....@ extended....@.............@ 122 GiB..@ -........@ ./dev/sda5...@ ext4........@ / ..........@ 37 GiB...@ 6 GiB....@ .unallocated @ unallocated @.............@ 7 GiB....@ - ...... @ ./dev/sda6 ..@ ext4........@ /home.......@ 77 GiB...@ 32 GiB...@ .unallocated @ unallocated @.............@ 65 GiB...@ - .......@ /dev/sda3...@ linux-swap..@.............@ 7 GiB....@ - .......@ My plan is to: - boot to ubuntu from USB ISO - change sda1 to NTFS - install W7 to sda1 - use the "Master Boot Record repair" utility to configure dual boot so I can see my original ubuntu installation as well as W7. Have I missed something? I'm concerned as to what the 776MB is that will be overwritten by the change to NTFS. It seems large for just the MBR? Would also appreciate it if anyone can explain what sda5 and 6 are being used for? Is sda5 Ubuntu and sda6 my data? Thanks in advance.

    Read the article

  • options for deploying application

    - by terence
    I've created a simple web application, a self-contained tool with a user system. I host it publicly for everyone to use, but I've gotten some requests to allow companies to host the entire application privately on their internal systems. I have no idea what I'm doing - I have no experience with deployment or server stuff. I'm just some person who learned enough JS and PHP to make a tool for my own needs. The application runs with Apache, MySQL, and PHP. What's the best way to package my application to let others run it privately? I'm assuming there's better options than just sending them all the source code. I'd like to find a solution that is: Does not require support to set up (I'm just a single developer without much free time) Easy to configure Easy to update Does there exist some one-size-fits all thing that I can give to someone, they can install it, and bam, now when they go to http://myapplication/ on their intranet, it works? Thanks for your help.

    Read the article

  • Upload large database SQL file

    - by Devy
    I've a database of more than 20Gb of size on my hard disk. What is the best way to upload it with the least (money) load possible on the server? - I'm on Windows 7. - I have FTP and SSH access on the server. I avoid using FTP because my connection cuts off a lot, I can't imagine I re-upload again the file after failing on 99%. I found some tools that split the large .sql file to small .sql files, but they didn't mention how to gather these files again into one file. Another way is to archive the big .sql file to .rar with -v option, upload them through FTP then unpack them. But unpacking will also cost, right? I know it will cost in any cases, but any best practice will be strongly appreciated.

    Read the article

  • HP DV5T laptop problems using Adobe in multiple monitors

    - by Rolnik
    I have a dualview arrangement of two monitors that works just fine for 90% of my applications. However, any time I open Adobe Acrobat Reader to the larger monitor, it fails to permit enlarging of the window to fill the entire screen. In other words, it seems Adobe crops the PDF view at about the size of the smaller of the two monitors, despite Acrobat being present in the larger monitor. I'm using nVideo graphics on the laptop, and connecting to the larger monitor through the docking port of a docking station. Any ideas how to get the Acrobat to be aware that it has more screen real-estate and use it fully?

    Read the article

  • Dim (NEARLY blank) laptop screen, secondary screen works - why?

    - by LIttle Ancient Forest Kami
    My laptop screen is (almost) black while my secondary screen is fine. I believe it to be backlight / brightness related. Problem description it starts when I start the laptop system loads and works fine, just screen has problems I can see the screen though very faintly / dimly - it's hard to see anything which ain't very white e.g. starting screen has big Thinkpad logo in white, large font - I can see it, though very dimly second screen works very well Official backligtht debugging: using acpi setting as prescribed there for Thinkpads didn't help I can see an entry in /sys/class/backlight/ and it changes when I press hotkeys for brightness (current backlight power for instance goes up or down) acpi-off didn't helpm neither did acpi_backlight=vendor Hardware data Laptop is Thinkpad Edge with glossy screen. 4 processors, 2 cores, exemplary CPU data from cat /proc/cpuinfo reports Genuine Intel i5 (M 480 @ 2.67GHz). OS is Ubuntu Lucid, 10.04 LTS, 64-bit, with Linux generic kernel (2.6.32-44) and GNOME 2.32.2 (though I doubt there lies the problem). $ lspci | grep VGA 01:00.0 VGA compatible controller: ATI Technologies Inc M92 [Mobility Radeon HD 4500 Series] $ lshw -C display *-display description: VGA compatible controller product: M92 [Mobility Radeon HD 4500 Series] vendor: ATI Technologies Inc physical id: 0 bus info: pci@0000:01:00.0 version: 00 width: 32 bits clock: 33MHz capabilities: pm pciexpress msi bus_master cap_list rom configuration: driver=radeon latency=0 resources: irq:33 memory:c0000000-dfffffff(prefetchable) ioport:2000(size=256) memory:f0300000-f030ffff memory:f0320000-f033ffff(prefetchable) Driver I was NOT running any proprietary drivers, just checked with "Hardware drivers". There is one for ATI that is suggested there, though I didn't need it so far. UPDATE: changing the driver to proprietary one (ATI/AMD FGLRX) didn't help. Tried and failed Resetting / running on power or battery / charging / getting rid of static electricity / warming up *doesn't help* This is NOT a blank-screen problem, at least it isn't following official Ubuntu black-screen diagnostics - I can see my screen, though barely. What I will try next: - check last updates I've made - IIRC I am running on nomodeset already, but will verify this Any ideas how to proceed best? What is most probable cause?

    Read the article

< Previous Page | 443 444 445 446 447 448 449 450 451 452 453 454  | Next Page >