Search Results

Search found 21028 results on 842 pages for 'single player'.

Page 521/842 | < Previous Page | 517 518 519 520 521 522 523 524 525 526 527 528  | Next Page >

  • EXC_BAD_ACCESS error when box2d joint is destroyed

    - by colilo
    When I destroy the weldJoint in the update method (see below) I get an EXC_BAD_ACCESS error pointing to the line world->DestroyJoint(weldJoint); in the update method below: -(void) update: (ccTime) dt { int32 velocityIterations = 8; int32 positionIterations = 1; // Instruct the world to perform a single step of simulation. It is // generally best to keep the time step and iterations fixed. world->Step(dt, velocityIterations, positionIterations); // using the iterator pos over the set std::set<BodyPair *>::iterator pos; for(pos = bodiesForJoints.begin(); pos != bodiesForJoints.end(); ++pos) { b2WeldJointDef weldJointDef; BodyPair *bodyPair = *pos; b2Body *bodyA = bodyPair->bodyA; b2Body *bodyB = bodyPair->bodyB; weldJointDef.Initialize(bodyA, bodyB, bodyA->GetWorldCenter()); weldJointDef.collideConnected = false; weldJoint = (b2WeldJoint*) world->CreateJoint(&weldJointDef); // Free the structure we allocated earlier. free(bodyPair); // Remove the entry from the set. bodiesForJoints.erase(pos); } for(b2Body *b = world->GetBodyList(); b; b=b->GetNext()) { if (b->GetUserData() != NULL) { CCSprite *mainSprite = (CCSprite*)b->GetUserData(); if (mainSprite.tag == 1) { mainSprite.position = CGPointMake( b->GetPosition().x * PTM_RATIO, b->GetPosition().y * PTM_RATIO); CGPoint mainSpritePosition = mainSprite.position; if (mainSprite.isMoved) { world->DestroyJoint(weldJoint); } } } } } In the HelloWorldLayer.h I set the weldJoint with the assign property. Am I destroying the joint in the wrong way? I would really appreciate any help. Thanks

    Read the article

  • At what point does caching become necessary for a web application?

    - by Zaemz
    I'm considering the architecture for a web application. It's going to be a single page application that updates itself whenever the user selects different information on several forms that are available that are on the page. I was thinking that it shouldn't be good to rely on the user's browser to correctly interpret the information and update the view, so I'll send the user's choices to the server, and then get the data, send it back to the browser, and update the view. There's a table with 10,000 or so rows in a MySQL database that's going to be accessed pretty often, like once every 5-30 seconds for each user. I'm expecting 200-300 concurrent users at one time. I've read that a well designed relational database with simple queries are nothing for a RDBMS to handle, really, but I would still like to keep things quick for the client. Should this even be a concern for me at the moment? At what point would it be helpful to start using a separate caching service like Memcached or Redis, or would it even be necessary? I know that MySQL caches popular queries and the results, would this suffice?

    Read the article

  • Set one background to stretch across multi-monitor display?

    - by John Isaacks
    I have two monitors and my coworker has three. We were wondering if there was a way to set a single background image to stretch across all the monitors. Right now it's the same image but repeated, so it looks like the same image three times in a row. I would like it to be so that the three images/screens combine to show one image if that makes sense. If I can't set one image to stretch across the screens like that, is there a way to set a different background for each screen? Edit: I just wanted to clarify: I am not talking about one image stretching from his monitors to mine, just stretching across our own monitors.

    Read the article

  • Windows failover cluster - virtual node with multiple Client Access Names

    - by mclaassen
    I recently encountered a customer environment in which they had failover cluster where one of the virtual nodes had more than one 'Client Access Name' (i.e. more than one IP address and DNS name for the single virtual node). Long story short we had to modify our software to deal with this situation, but we want to recreate the situation in house to test it before releasing. I have been unable to locate any information about how or why you would end up with a virtual node that has more than one access name. Does anyone know how I can set up a Windows failover cluster where a virtual node has more than one access name/IP?

    Read the article

  • How do I skip the Keep/Cancel dialog for downloads in Google Chrome?

    - by NoCatharsis
    This question is specifically for PDF files, because I work with them all day and have to download 20+ at a time. I use an extension called Linkclump which selects all links at one time. Since I disabled the Chrome PDF Viewer, these now automatically download (which I intended). The only problem is I get 20+ dialogs across the bottom of my screen that ask me if I want to Keep or Cancel my downloads. Seriously? I have to click Keep on every single one of them. This did not happen before in Chrome, the change was added a few versions ago I believe. Oh, and I'm using Chrome Dev v14.0.803.0.

    Read the article

  • What security changes are necessary when connecting DSL modem directly to PC instead of router?

    - by Mike B
    Windows XP I have a user with a single PC that was connected to the internet via a standard home router. The router is now having hardware-related issues and to save money, they're considering connecting the PC directly to the DSL modem since they don't need to share the internet connection or need wireless functionality. If they decide to do that, I'm concerned that this will introduce additional security concerns. Is the Windows Firewall sufficient and Microsoft Security Essentials sufficient for protecting a computer directly connected to a DSL Modem? Or is other security software needed here? Ideally, I'd like to avoid having third-party firewall software constantly bringing up alerts and asking them to approve everything. Also, just to clarify, their use cases are just internet browsing and email.

    Read the article

  • How can I get all my itunes files from multiple computers onto my new PC?

    - by jasondavis
    Managing my itunes files (music,video,tv shows, iphone apps) is a really hard task and I have not had much luck yet. If I had a single PC for the rest of my life with the same hard drive, it would work great but infortunately it cannot be that simple. I just build a new PC as my old hard drive crashed with all my itunes purchased songs, video, etc. So is there anyway to get itunes to recognize that I own all them and let me re-download everything? As I see it so far, it knows what I own by my itunes account login (it knows if I purchased a file or not ever) but does not seem to let you just re-download everything to a new PC or new hard drive. I realize I should of had backups but I didn't. My backup plan is a work in process and it is getting much better but that doesnt help with my situation at the moment. So I am asking for any help, ideas of how to retrieve all my previous itunes purchases?

    Read the article

  • How to avoid being fooled

    - by Dacav
    I'm a big fan of the OpenSource development model, and I think that sharing information, knowledge and ideas is the best way of working in software development. Still I think that being hired for proprietary software development must not be demonized. Of course, as there's a no reward in terms of sharing there must be a bigger reward in terms of money (i.e. I surrender all my rights for cash). It may happen that one gets hired piecework, for a single project: in this case one is more vulnerable to dishonest employers. This didn't happen to me personally, but some friends of mine had bad experiences, and lost a lot of time without being fairly retributed. Of course a contract should protect both parts. But contracts can be very generic in the specification. Software is not anything but a palpable good! Besides I don't think that contracts can distinguish between a well written software and a poorly written one. Note also that, in this (nasty) spirit, it's also likely that the employer cannot trust the employee! So also the employer should be protected by a dishonest employee). My question is the following: Which is, in your opinion, a good way of avoiding this kind of situation from the technical perspective?

    Read the article

  • Mac OS X multi-user thin client server (terminal server)?

    - by username
    Is there any solution out there to turn a Mac into a true multi-user thin client server? I'd like to set up a few cheap PCs with access to a couple accounts using something like VNC, but it isn't economical to buy a new server for each user or a new license for virtualized OS X Server for each user. I'm fully aware that OS X Server lets you set up users with "network home folders," and I know there's also VNC built into Mac OS X. Neither of these fit the bill (the former requires a thick client, and the latter is single-user only) UPDATE: yay, Lion! http://www.9to5mac.com/54102/10-7-lion-allows-multi-user-remote-computing

    Read the article

  • Need to hookup HP dv7-3085dx with NVIDIA GeForce GT 230M to my Dell 30 inch LCD 3007WFP at max resolution

    - by user14660
    I recently bought an HP laptop (dv7-3085dx) which is supposed to have a pretty good video card (NVIDIA GeForce GT 230M). The card is supposed to output a max resolution of 2560x1600 which is also the max resolution of my monitor. I've now bought an HDMI to dual link DVI cable - this is after Best Buy's 70 dollar HDMI to DVI (perhaps it was 'single' link?) didn't give me the best resolution. In Windows 7, when I try to set the max resolution for my 30" monitor, I only get 1280x800, which is absurd. The monitor is great, I love the laptop and the video card supposedly supports such resolutions. I therefore can't figure out why I'm not getting a better resolution. When I "detect" my monitor in Windows 7, it is shown correctly as a DELL 3007WFP!

    Read the article

  • tomcat processParameters complains about "invalid chunk ignored"

    - by cgicgi
    I am hosting a software system running under tomcat for quite a number of customers. Some of these send invalid URLs as request. These URLs may contain "&=" or "&&", which is not within the http specs. Now my tomcat complains about the following: "08.09.2010 12:36:04 org.apache.tomcat.util.http.Parameters processParameters WARNING: Parameters: Invalid chunk '' ignored." It is no problem, as is doesn't affect the operation in any way. Only problem ist that the tomcat/logs/catalina.out is growing with every single request. In the net you can find suggestions like: - Fix your URLs (which I can't, as it is the customers who send them) - Raise tomcats log level to ERROR (which I don't want to do, as it would suppress INFO like "INFO: Reloading context [/ContextName]" and other stuff you want to know. - Redirect the log to the application log (which won't solve the problem, as the message will flood just another log) Does anyone know how to solve the problem at its ROOT, which means: Tell tomcat not to complain about invalid request parameters any longer

    Read the article

  • How to maximize parallel download from S3

    - by StCee
    I got a lot of images to load from Amazon S3 on a single page, and sometimes it takes quite some time to load all the images. I heard that splitting the images to load from different sub-domains would help parallel downloads, however what is the actual implementation on that? While it is easy to split for sub-domains like static,image, etc; Should I make like 10 sub-domains (image1, image2...) to load say 100 images? Or is there some clever ways to do? (By the way I am considering using memcache to cache the S3images; I am not sure if it is possible. I would be grateful for any further comments. Thanks a lot!

    Read the article

  • Make taskbar buttons fixed on Windows 7

    - by Tchypp
    Sometimes the application buttons on the Windows 7 taskbar change position by themselves. For example, if I have two windows of the same application opened, the first opened window should be the first button on the taskbar (to the left of the other) and the second opened window should be the second button on the taskbar (to the right of the other), but they switch places sometimes. I don't like this. How can I fix / configure this not to get rearranged? I am not using Aero, and the buttons of the same application are not grouped under a single button. !

    Read the article

  • Why doesn't my cron.d per minute job run?

    - by Travis Griggs
    I have thrown a bunch of darts trying to get a python script of mine to execute every minute. So I thought I'd simplify it to just do the "simplest thing that could could possibly work" once per minute (I'm running debian/testing). I created a single line file in /etc/cron.d/perminute: * * * * * /bin/touch /home/me/ding_dong It's owned by root, and executable (not sure if either of those matter). And then I did: sudo service cron reload And then sit back and start running ls -ltr again and again in my home directory (/home/me). But my ding_dong file never shows up. I know if I do a sudo /bin/touch /home/me/ding_dong, it shows up right away. Obviously missing something stupid here.

    Read the article

  • Ubuntu 12.04 "stuttering"

    - by Totte
    When I log in to the desktop Ubuntu seems to "stutter". Doing pretty much anything causes the cursor to freeze for a few seconds, and the rest of the screen updates something like every 15 seconds (during these 15 seconds I can still move the cursor). For example, opening the home folder I first see no effect, followed by a semi-transparent home folder, and finally the folder as it should be (this can take like 30 seconds); no animation, just three individual frames. Also, I sometimes don't see a window for a program, even though the program in question is supposed to be "up". Since this is my first experience of Linux, I confess I have no idea where to start - searching the web I only found problems with complete freezes/crashes requiring rebooting, as opposed to this "stuttering". I installed Ubuntu 12.04 32bit from a bootable USB stick a couple of days ago, right after a fresh Vista 32bit install. Thinking the stuttering might stop if I updated Ubuntu, I managed to install approx 270 updates through the Update Manager, but the stuttering was still there. Apart from the updates, I haven't installed any software other than what was included on the bootable USB stick. Vista and Ubuntu are on a single Crucial m4 SSD, in a dual-boot setup, with GRUB. CPU: Intel Core 2 Quad Q6600 GPU: Nvidia GeForce 8800 GTX (with the recommended proprietary drivers) RAM: 2GB DDR2

    Read the article

  • Do control groups improve system performances?

    - by qdii
    According to this website, enabling cgroups in the kernel can boost performances by sharing resources in a better way. In particular, the conclusion states that:  Nevertheless, with a little trial and error, cgroups can help you improve the efficiency of your systems’ resource usage and avoid downtime due to overusage of a single service. Kernel seeds, however, recommend to deactivate them altogether. They say: Consider these [kernel] settings poison. They remain nothing but system slow-downs. They are all off by default [in the proposed kernel config file]. Who should I trust?

    Read the article

  • Why are browsers so heavy?

    - by Kaivosukeltaja
    Back in 1998 I had a computer with 233MHz Pentium MMX CPU and a GFX card with no 3D acceleration. It was able to run games like Quake II at a decent FPS rate. My current computer has tons more performance and a mid-class GPU, yet struggles to reach 20 FPS when rendering a single model inside a skybox with WebGL. Even regular pages with lots of 2D CSS animations bring many modern computers to their metaphorical knees. As a web developer I understand there's a lot going on in a web page but not what makes it that heavy. Modern browsers compile JavaScript to CPU native machine code before running it and rendering into a canvas element shouldn't trigger DOM rebuilds so theoretically it should be a lot faster than it is. What am I missing here and is it possible to avoid or minimize whatever is making the browsers slow to build more efficient websites?

    Read the article

  • CPU Configuration Issue for 2 Servers (Server 2008 R2)

    - by Bill Moreland
    I have 2 servers running the exact same Classic ASP code with Access DBs (yes, not ideal, but it is what it is, for now). 1) Xeon 5520 @ 2.27 GHz (6 GB Memory) 2) Xeon E5-2620 @ 2.00 GHz (2 processors, 32 GB Memory) For most pages the newer E5-2620 processes the pages between 10-15% faster. On pages requiring heavy and/or multiple complicated access stored procedures (queries) the older 5520 does a much better job. I believe the servers are configured nearly identically. My question: is it possible that the newer, multi-processor server is not as good at handling Classic ASP as the older single processor? Is there a configuration difference that needs to be in place that I'm missing since I'm shooting for identical implementations?

    Read the article

  • How to report spam to blacklists

    - by hayalci
    Is there a central place to report spam to various blacklists ? I regularly report to spamcop, but I do not see the addresses I reported as listed. (I guess nobody else bothers with my regular spammer. After being frustrated with bayes and spammcop, I blocked its /24 subnet) Spamcop is only a single service. I want to make the spammer known to a large number of services, and hopefully blocked by many of them. I looked for some other blacklists to report to, but the ones I looked do not consider user submissions (or they hide it well)

    Read the article

  • Does Microsoft offer a corporate IM/collaboration tool similar to Campfire? My googlefu skills appear to be failing me today.

    - by user54266
    I mentioned to my boss that we should look into a single unified IM client that we could use and secure on a corporate level, and then suggested Campfire. We're a primarily Microsoft house so he suggested we use something that would better integrate with SharePoint and the other tools our end users use in house. However, I'm not aware of any Microsoft tool that does something like this. Obviously there is MSN Messenger but I think/hope he wasn't referring to that. Other than a product from 2005 I haven't been able to locate a Microsoft corporate IM tool...does anybody know what he may have been talking about?

    Read the article

  • Access denied for user who has full access to some files in their own folder

    - by steve02a
    I have a very similar case as this user: Access denied on some files on Win2008R2 DC share This is on a windows 2008 R2. The user has Win7 pro. The user has their own home folder on the server. Every file, except one, the user can read/write/modify at their own will. No problems - except this one file. She gets "access denied" I can open it (as domain admin). Another user can open it (because she's in the domain admin group). I did run the AccessEnum tool and the read/write permissions are all identical for all files. So, I can't explain why the user can't open this one single file. Out of all her files in sub-folders and such. No problems. This one file is causing a headache. What do you think could be wrong here?

    Read the article

  • How do I force an exchange database to become "active"?

    - by makerofthings7
    We had a catastrophic failure where all that remains is a single edb file. No backups. No log files. The database that remains is on the "passive" copy. The "active" copy is missing, but the server is active. The Exchange console reports that the edb file needs to be reseeded, however there is no source to reseed from. How do I make the "invalid" database file (missing logs) valid? How do I make exchange recognize this as a valid database to use as a primary?

    Read the article

  • Move already (i.e. not new) completed files in rtorrent

    - by Skippy le Grand Gourou
    I've been using rtorrent for a while with a single directory. Now I figured out it was possible to use different directories and even to move completed downloads elsewhere, so according to the rtorrent wiki I edited my .rtorrent.rc as follow : # Download directory directory = /Medias/torrents/ # Watching directories schedule = watch_directory_1,5,60,"load_start=/path/to/dl/dir1/*.torrent,d.set_custom1=/path/to/done/dir1" schedule = watch_directory_2,5,60,"load_start=/path/to/dl/dir2/*.torrent,d.set_custom1=/path/to/done/dir2" # On completion, move the torrent to the directory from custom1. system.method.set_key = event.download.finished,move_complete,"d.set_directory=$d.get_custom1= ;execute=mv,-u,$d.get_base_path=,$d.get_custom1=" It seems to work for new torrents. However I have a bunch of completed files I have already downloaded before to split directories, and for them it does not work : if I delete their file in the session directory, rtorrent will check the hash but won't move them, and if I move them by myself rtorrent won't see them and will try to redownload them. So how can I tell rtorrent either to move them or that they are in another directory ? Thanks.

    Read the article

  • Remote additional domain controllers

    - by user125248
    Is it possible to setup several additional domain controllers (ADC) at remote locations that are connected via medium bandwidth DSL (2-10 Mbit) WAN connections for a single domain (intranet.example.com)? And would it be a good idea? We have five sites and would like to have extremely high availability if any of the site were to lose their Internet connection. However each site is very small, and all are over a fairly small geographical area within the same region, so it would seem strange to have a PDC for each of the sites. If it were possible to have an ADC for each site, would the clients use the ADC or just use the PDC if it's available to them?

    Read the article

  • What are the pros & cons of these MySQL engines for OLTP -- XtraDB, PBXT, or TokuDB?

    - by Continuation
    I'm working on a social website with an approximate read/write split of 90/10. Trying to decide on a MySQL engine. The ones I'm interested in are: XtraDB PBXT TokuDB What are the pros and cons of them for my use case? A few specific questions: PBXT uses log-based structure that avoids double-writes. It sounds very elegant, but the benchmark I've seen doesn't show any/much advantages over XtraDB. Do you have any experience with PBXT/XtraDB you can share? TokuDB sounds VERY interesting. But all the benchmarks I've seen are about single-threaded bulk inserts - inserting 100M rows for example. that's not very relevant for OLTP. What about its performance with large number of concurrent threads writing and reading at the same time? Anyone has tried that?

    Read the article

< Previous Page | 517 518 519 520 521 522 523 524 525 526 527 528  | Next Page >