Search Results

Search found 19579 results on 784 pages for 'image segmentation'.

Page 599/784 | < Previous Page | 595 596 597 598 599 600 601 602 603 604 605 606  | Next Page >

  • 7.0.1 and Photoshop openGL requirements

    - by mcgd
    Hi, I'm running Workstation 7.0.1 on a 64-bit Linux host, with a Vista guest, and the Aero theme works a treat - full marks for getting that working! However, when running Photoshop there are some operations (like interactive image rotation) that seem to require graphics support that VMWare doesn't have - attempting to do these operations returns an error that the functionality "only works with OpenGL enabled windows". It seems that it needs OpenGL 2.0 and Shader Model 3.0 which I gather are supported under Workstation 7 - do I need to put in additional settings of some kind? I've updated to the right version of VMWare Tools, and the video driver is listed as "VMWare SVGA 3D (Microsoft Corporation - WDDM)". I see that in 6.5 there was an option to switch between the beta WDDM driver and the SVGAII driver, but I assume the features of both of these would have been merged into the standard driver in 7.0? Thanks -Matthew

    Read the article

  • Can't boot Windows after installing Linux

    - by user4035
    I have a partition /dev/sdb1, where my old Windows XP resides. All the files are there intact and I can see them, mounting the disk from Linux. Linux is on /dev/sdb2. But when I choose Windows in LILO prompt, it doesn't load. I have the following lilo.conf: boot = /dev/sdb # Linux bootable partition config begins image = /boot/vmlinuz root = /dev/sdb2 label = Linux read-only # Partitions should be mounted read-only for checking # Linux bootable partition config ends # Windows bootable partition config begins other = /dev/sdb1 label = Windows table = /dev/sdb # Windows bootable partition config ends What can be wrong?

    Read the article

  • What are the current options to encrypted a partition on mac os x ?

    - by symbion
    I recently got my laptop stolen with some sensitive informations on it (personal source code, bank details in a secure file, passwords, etc) and I learnt the lesson: encrypt your sensitive data. Now, I am wondering what are the options to encrypt a partition (not an encrypt disk image) ? Aim: The aim is to prevent anyone (except me) to access those data. Requirement 0: The software must be able to encrypt non system partition. Requirement 1: Plausible deniability is required but preventing cold boot attack is however not an absolute requirement (I am not famous enough or have sensitive enough info to have this kind of requirement). Requirement 2 : Software taking advantage of AES hardware encryption are very welcome as I intent to get a Macbook Pro with i7 CPU (with AES-NI enabled instructions). I will have avirtual machine running in the encrypted partition. Requirement 3 : Free or reasonably cheap. Requirement 4 : Software must run on Mac OS X Snow Leopard or Lion. So far, TrueCrypt is the only option I have found. Regards,

    Read the article

  • How do I add a boot from cd option to yaboot?

    - by Sergiu
    So I'm dual-booting Ubuntu 12.04.1 on my iMac G5 powepc alongside Mac OS X and I want to add a boot cd option to yaboot because I'm trying to boot a scratched Mac OS X installation DVD that takes a while to read and the frst bootstrap moves on too fast. How do I edit the timeout for the first bootstrap anyways? So, my main question is, how do I add a cd booting option to yaboot and then, how doI boot it? The devalias from OpenFrmware tells me that 1 have 2 cd-rom instaled, on is /ht/pci@3/ata-6/disk@0 and the other on ends with a 1 instead of a zero. These are the contents of my yaboot.conf file: yaboot.conf generated by the Ubuntu installer run: "man yaboot.conf" for details. Do not make changes until you have!! see also: /usr/share/doc/yaboot/examples for example configurations. For a dual-boot menu, add one or more of: bsd=/dev/hdaX, macos=/dev/hdaY, macosx=/dev/hdaZ boot="/dev/disk/by-id/scsi-SATA_ST3160023AS_5MT1GCWA-part2" device=/ht@0,f2000000/pci@3/k2-sata-root@c/@0/@0 partition=4 root="UUID=798a048f-ee48-49e0-bba3-111aed8dee04" timeout=12000 install=/usr/lib/yaboot/yaboot magicboot=/usr/lib/yaboot/ofboot enablecdboot macosx="/dev/disk/by-id/scsi-SATA_ST3160023AS_5MT1GCWA-part3" image=/boot/vmlinux label=Linux read-only initrd=/boot/initrd.img append="quiet splash" What do I add here so that yaboot will boot from my cd in like 3 minutes after startup? Thanks!

    Read the article

  • Exceptions from automongobackup, yet script completes

    - by chakram88
    I am using automongobackup to, well, automate the backups of mongodb. output from the script (to STDERR) has the following exceptions (but the backup completes, and the dump files are created) ###### WARNING ###### STDERR written to during mongodump execution. The backup probably succeeded, as mongodump sometimes writes to STDERR, but you may wish to scan the error log below: exception: connect failed exception: connect failed exception: connect failed exception: connect failed exception: HostAndPort: bad port # exception: connect failed exception: connect failed exception: connect failed exception: connect failed exception: connect failed exception: connect failed I know that the Host & Port are correct. If I run mongodump --host=127.0.0.1:27017 --journal (which is the effective command from automongobackup based on the options set and my reading of the src code) everything runs clean without any error reporting and the dump files are created as expected. Why would automongobackup report connection errors, even tho it does create the dump files, yet a straight call to mongodump does not? Debian 6.0 Lenny (from Linode image: Latest 3.2 (3.2.1-x86_64-linode23)) AutoMongoBackup VER 0.9 mongodb v 2.0.2

    Read the article

  • Odd behavior of permissions and icons

    - by Urban
    After some virus had infected my Windows 7, I did a complete format and re-installed the OS. I was just installing applications and copying back some data when I noticed some shortcuts changing their icons to something I couldnt recognize (Yellow icons in the image below). Also, a few exe files which previously did not ask for User permission, are now asking for it. Wondering if this is an icon cache problem, I cleared the cache by deleting the IconCache.db in AppData/Local but the problem still persists. Although I did a full system scan with MS Security Essentials, I'm not sure if this is another virus or some other problem. I would appreciate any suggestions you might have. Edit: Now even Firefox needs permissions to launch. It's icon hasnt changed, but it's got the 'shield' overlay on it like the other yellow icons.

    Read the article

  • How do I PXE Boot Only when I want it to without user intervention?

    - by troz123
    I would like to setup a PXE environment where I can re-image machines remotely without any user intervention. Only problem is when the re-imaging is completed it will do the re-imaging again and again and again. If I remove the MAC address file then I just get a error saying it can't find the MAC address file and the system stops. I also tried turning off the TFTP server and I get a error stating can't find TFTPD server. How can I make client machines only PXE boot once and after the re-imaging it will boot into Windows and everything is happy? And only PXE boot when I want it too... I'm using TFTPD32 to serve the files. I'm using a Windows 2003 DHCP server that points to pxelinux.0...

    Read the article

  • Replicate a big, dense Windows volume over a WAN -- too big for DFS-R

    - by Jesse
    I've got a server with a LOT of small files -- many millions files, and over 1.5 TB of data. I need a decent backup strategy. Any filesystem-based backup takes too long -- just enumerating which files need to be copied takes a day. Acronis can do a disk image in 24 hours, but fails when it tries to do a differential backup the next day. DFS-R won't replicate a volume with this many files. I'm starting to look at Double Take, which seems to be able to do continuous replication. Are there other solutions that can do continuous replication at a block or sector level -- not file-by-file over a WAN?

    Read the article

  • Pass User Data to AWS client

    - by bearrito
    Has anyone successful passed user data to the AWS CLI ? I have tried various incantations of the following but it does not work. Docs say string must be base64 encoded : http://docs.aws.amazon.com/cli/latest/reference/ec2/run-instances.html The instance logs never indicate the script is executed and chef is installed. aws ec2 run-instances --image-id ami-a73264ce --count 1 --instance-type t1.micro --key-name scrubbed --iam-instance-profile Arn=arn:aws:iam::scrubbed:instance-profile/scrubbed --user-data $(base64 chef_user_data.sh --wrap=0) chef_user_data.sh #!/bin/bash curl -L https://www.opscode.com/chef/install.sh | sudo bash

    Read the article

  • Bad password when logging in via Remote Desktop

    - by Shiraz Bhaiji
    I have just setup a windows 2003 server running in hyper-V. When I log in via Hyper-V it works fine. When I try to login via remote desktop I get a bad user name or password error. I have: Disabled the firewall Enabled remote connections Before I did the above I did not get to the login screen. I can ping the ip address of the image. I am trying to login as Administrator. Anyone have an idea about how to fix this? Edit I have now managed to fix this by changing the password to something without special characters. Strange thing is the I could not see any problems with the passord, had tried typing it into the user name field to check it.

    Read the article

  • Artificial Intelligence ... how to make an object roam freely/avoid other objects, and model consciousness? [on hold]

    - by help bonafide pigeons
    Say a simple free roam battle scene in which a player runs around freely and engages in battle with other enemies/objects, as shown below: The dragon/dinosaur (or whatever that thing I drew appears to be) will, by some measure, try and avoid attacks so it is modeled to appear to have a conscious desire to avoid pain. My question is ... since this is very complex, many possible strategies for solving this, algorithms, etc., what is the basic idea behind how this would be accomplished in any sort? Like, we can assume the enemy in the picture is not just going to aimlessly hop around and avoid, but freely be modeled to behave as if it were really exploring/fighting. For the best example I can give, witness the behavior of the enemies in Final Fantasy 12 in this video: http://www.youtube.com/watch?v=mO0TkmhiQ6w How do the pros, or how would anyone attempt solve/implement this? PS: I have tried several times to give an image the "illusion" that is has a conciousness, but aside from emulating a real animal's consciousness in complete, I fall short and get choppy moving images that follow predictable patterns, error-prone movements, and the worst imaginable scenario of a battle engagement.

    Read the article

  • Mod_rewrite on Alias directory

    - by crunchline
    I am moving from a WIMP setup to a WAMP setup. My images are on a separate drive. I need a default image served, when a file is not found. Currently it returns a 404 on images not found in the /images directory, and the 404.jpg is displayed for files not found on all the other directories. I tried changing [L] to [PT] but that did not do anything. in httpd.conf Alias /images "D:/images" in .htaccess RewriteCond %{REQUEST_FILENAME} !-s RewriteRule ^.*\.(gif|jpg|png)$ /404.jpg [L]

    Read the article

  • Higher Performance With Spritesheets Than With Rotating Using C# and XNA 4.0?

    - by Manuel Maier
    I would like to know what the performance difference is between using multiple sprites in one file (sprite sheets) to draw a game-character being able to face in 4 directions and using one sprite per file but rotating that character to my needs. I am aware that the sprite sheet method restricts the character to only be able to look into predefined directions, whereas the rotation method would give the character the freedom of "looking everywhere". Here's an example of what I am doing: Single Sprite Method Assuming I have a 64x64 texture that points north. So I do the following if I wanted it to point east: spriteBatch.Draw( _sampleTexture, new Rectangle(200, 100, 64, 64), null, Color.White, (float)(Math.PI / 2), Vector2.Zero, SpriteEffects.None, 0); Multiple Sprite Method Now I got a sprite sheet (128x128) where the top-left 64x64 section contains a sprite pointing north, top-right 64x64 section points east, and so forth. And to make it point east, i do the following: spriteBatch.Draw( _sampleSpritesheet, new Rectangle(400, 100, 64, 64), new Rectangle(64, 0, 64, 64), Color.White); So which of these methods is using less CPU-time and what are the pro's and con's? Is .NET/XNA optimizing this in any way (e.g. it notices that the same call was done last frame and then just uses an already rendered/rotated image thats still in memory)?

    Read the article

  • How can I make a non-destructive copy of a (NTFS) partition?

    - by violet313
    I want to recover some deleted files from a healthy NTFS partition on an undamaged hard-disk. In order to leave the partition undisturbed, i plan to use dd to clone the partition to a raw image file & then attempt recovery from that mounted clone. Will dd if=/dev/sd<xn> of=/path/to/output.img perform a non-destructive copy ? Is attempting a restore from a clone using dd the best approach? [edit, wrt Deltiks answer, i need to be a bit clearer about what i'm asking] eg: are there some s/w that can do something more with the original sectors ? eg: if it was a damaged hard-disk i am aware that any kind of read is potentially destructive. but assuming my disk head is not going to suddenly spaz out etc, am i reducing my chances of a successful recovery (at any cost) by using an apparently non-destructive single read of my undamaged hard-disk. (btw: i am planning on using ntfsundelete & testdisk for recovery)

    Read the article

  • Browser Item Caching and URLs

    - by Damon
    Ultimately you want the browser to cache things like Flash components, Silverlight XAP files, and images to avoid users having to download them each time they hit a page.  But during development it's very useful to NOT have things cached so you are always looking at the most up-to-date file.  You can always turn off caching on your browser, but if you use your browser for daily browsing then its not the greatest option.  To avoid caching we would always just slap a randomly generated GUID to the back of the URL of any items we didn't want to cache (e.g. http://someserver.com/images/image.png?15f073f5-45fc-47b2-993b-fbaa781b926d).  It worked well, but you had to remember to remove the random GUID when it went to production. However, on a GimmalSoft project we recently implemented someone showed me a better way that didn't need to be removed from production code - just slap the last modified date of the file on the end of the URL (or something generated from the modification date).  This was kind of genius approach because it gives you the best of both world.  If you modify the file, the browser goes out and gets the newest version.  If you don't modify the file, it has the cached copy.  Very helpful!  The only down side is that you do have to read the modification date from the file, which does technically take some time.

    Read the article

  • Browser Item Caching and URLs

    - by Damon Armstrong
    Ultimately you want the browser to cache things like Flash components, Silverlight XAP files, and images to avoid users having to download them each time they hit a page.  But during development it’s very useful to NOT have things cached so you are always looking at the most up-to-date file.  You can always turn off caching on your browser, but if you use your browser for daily browsing then its not the greatest option.  To avoid caching we would always just slap a randomly generated GUID to the back of the URL of any items we didn’t want to cache (e.g. http://someserver.com/images/image.png?15f073f5-45fc-47b2-993b-fbaa781b926d).  It worked well, but you had to remember to remove the random GUID when it went to production. However, on a GimmalSoft project we recently implemented someone showed me a better way that didn’t need to be removed from production code – just slap the last modified date of the file on the end of the URL (or something generated from the modification date).  This was kind of genius approach because it gives you the best of both world.  If you modify the file, the browser goes out and gets the newest version.  If you don’t modify the file, it has the cached copy.  Very helpful!  The only down side is that you do have to read the modification date from the file, which does technically take some time.

    Read the article

  • Long domain lookup on .dev domain inside vmware

    - by skelle
    I'm developing on my macbook and normally I have a local running webserver which just works finde. Now I have to use a vmware image where the webserver is running. I set up everything and my dev site is running under site.dev inside vmware. I can connect to the webserver but EVERY request takes a very long time. I already red that this is related with iIPv6 and the way OSX handles /etc/hosts. There I added 192.168.155.42 site.dev and I already did this (Resolving to virtual host very slow on Mac OS X Lion) but my lookup still takes ~30seconds on every request. What can I do to fix this issue?

    Read the article

  • What are these completely random images that pop up when using Recuva?

    - by Qubert
    I just used Recuva to get back some photos I accidentally deleted on C:, but I also noticed there's a bunch of COMPLETELY random images in the list of recoverable files I can choose from. An example would be, There are images with names like John_Doe.jpg with a file path somewhere out of C:/Windows/Assembly/ And some of these things don't even make sense, there's one called componentsSpriteImage with the path C:/?/Images that shows a "poster" from the TV series Gold Rush. I am completely lost on this.. I know the OS is on C:, but I can't really think of a reason why Microsoft would leave a image hanging around from Gold Rush.

    Read the article

  • using modrewrite to change http to https? (not redirect)

    - by PaulHanak
    This might sound a little crazy, but bare with me. I basically have an include file, lets say inc-navigation.html, that has absolute paths (http://www.pathtoimage.com/image.com) that are on EVERY PAGE. Well, using SSL, I can't use that same include file because it is not referencing https:// What a pain! SO, I was maybe thinking of using htaccess to do a url rewrite of all references of HTTP to HTTPS when the browser requests a https page. Again, just to be clear, I don't want to "redirect", just "replace". So, I have this.... RewriteCond %{HTTPS} !=on RewriteRule ^http$ https but it doesn't seem to be working. I probably have the syntax wrong though. :) Of course, this is even if this type of thing is even possible!?

    Read the article

  • What is the cause of a vertical white line on an LCD monitor?

    - by Deviruchi D Devourer
    This just happened yesterday when I was browsing the internet. My AOC N941Sw monitor is suddenly showing a thin horizontal white line near the center of the screen. The white line distorts the images. I tried checking the cables for loose connection and dust particles but the white line is still here. There are times the entire image below the horizontal white line flickers and move up and down. When this happens, I just turn off the monitor then the flicker goes away but the white horizontal line is still there. After a few minutes, the same thing happen all over again. Is there a way to fix this?

    Read the article

  • Dual displays not working - NVidia - Ubuntu 12.04 - Second Monitor - Two Screens

    - by user75105
    Graphics Card: NVidia 460 GTX. Driver: NVIDIA accelerated graphics driver (version current) I have one DVI monitor, an old Dell LCD from 2005, and one VGA monitor, an Asus ML238H from 2010 whose HDMI port broke. The Asus is plugged into my graphics card's primary monitor slot and is the better monitor even though it is VGA but my computer defaults to the Dell. This happens when I boot as well; the loading screens, the motherboard brand image, etc. are all displayed on the Dell monitor until Windows loads. Then both monitors work. The same thing happened when I booted up Ubuntu 12.04 but I did not see the second monitor when the log-in screen popped up, nor did I when I logged in. I went to System Settings/Displays and my Asus monitor is not an option. I clicked Detect Displays and the Asus is not detected. I looked at the other questions regarding NVIDIA drivers and recalled my problems with Ubuntu a few years ago and decided to check the driver. I went to Additional Drivers to install the proprietary driver and it looks like it's installed and active but I'm still having this problem. There is another driver option, the post-release NVIDIA driver, but that does not fix the problem either. Also, under System Details/Graphics the graphics device is listed as Unknown, which might indicate that it is using an open source generic driver and not the proprietary NVidia driver. But under Additional Drivers it says that I am using the NVidia driver. Any help is appreciated.

    Read the article

  • Lost current user (shutdown/logoff) widget after upgrading to Ubuntu 10.04

    - by xyzman
    I've upgraded to 10.04 from 9.10. Everything went fine until I've rebooted. After reboot, I've received some message about status panel and being sleepy, dismissed it. However, I haven't seen the user widget on Gnome Panel since that. This image shows which widget I'm talking about. If some one has any ideas about how it's called properly, please share. http://img237.imageshack.us/img237/8274/sampleyu.jpg Anyway, the question is, how do I turn this panel (and ability to shut my system down without resorting to console) back?

    Read the article

  • Dual displays not working - NVidia - Ubuntu 12.4

    - by user75105
    Graphics Card: NVidia 460 GTX. Driver: NVIDIA accelerated graphics driver (version current) I have one DVI monitor, an old Dell LCD from 2005, and one VGA monitor, an Asus ML238H from 2010 whose HDMI port broke. The Asus is plugged into my graphics card's primary monitor slot and is the better monitor even though it is VGA but my computer defaults to the Dell. This happens when I boot as well; the loading screens, the motherboard brand image, etc. are all displayed on the Dell monitor until Windows loads. Then both monitors work. The same thing happened when I booted up Ubuntu 12.4 but I did not see the second monitor when the log-in screen popped up, nor did I when I logged in. I went to System Settings/Displays and my Asus monitor is not an option. I clicked Detect Displays and the Asus is not detected. I looked at the other questions regarding NVIDIA drivers and recalled my problems with Ubuntu a few years ago and decided to check the driver. I went to Additional Drivers to install the proprietary driver and it looks like it's installed and active but I'm still having this problem. There is another driver option, the post-release NVIDIA driver, but that does not fix the problem either. Also, under System Details/Graphics the graphics device is listed as Unknown, which might indicate that it is using an open source generic driver and not the proprietary NVidia driver. But under Additional Drivers it says that I am using the NVidia driver. Any help is appreciated.

    Read the article

  • How to partition my hard drive, quicker?

    - by Sam
    When I install Windows 7 on my hard drive, it makes three partitions. One with the OS itself, one with bootmgr inside (that is 100 MiB), and one with the factory image (all the crapware from HP). My final goal is to have the OS on a partition of 100 GiB and keep the rest (900 GiB) for storage. I thought it would be easy using gparted, but it is taking so long. It will take hours. There must a way to partition the drive before installing Windows. Yeah, because what I think makes the shrinking/moving of the partitions take so long is because they are not empty (am I wrong?).

    Read the article

  • Seagate 1TB 'EXPANSION' external hard drive, showing as BAD

    - by Dave Plews
    Hello! I have bough an external Seagate 1TB drive, which my XP system can see and will write files to, however when I use Partition Magic it describes the whole partition as BAD, and it is shown in yellow. All the options are greyed out. When I use Drive Image XML to try and copy drive to drive (I want a comlplete copy of my C drive including OS), I get an error message saying it can't lock the drive. The external drive is brand new and uses NTFS. Any ideas? Seagate 'support' haven't bothered getting back to me. Incidentally, I have another external drive (320GB), which uses FAT32 and Partition Magic sees that fine. I am doing a full format of the Seagate at the moment, to see if this helps. Thanks in advance!

    Read the article

< Previous Page | 595 596 597 598 599 600 601 602 603 604 605 606  | Next Page >