Search Results

Search found 17374 results on 695 pages for 'option strict'.

Page 437/695 | < Previous Page | 433 434 435 436 437 438 439 440 441 442 443 444  | Next Page >

  • Is there a PROPRIETARY driver (NVIDIA or ATI) that actually works with 12.10?

    - by DS13
    NOTE: I see many similar topics on this, but I've tried all their suggestions, and nothing has worked. THE MAIN DIFFERENCE SEEMS TO BE: I always get a black screen with a blinking cursor, while others seem to get through the boot-up and see distorted graphics or just their wallpaper. ISSUE: I do a clean install of Ubuntu 12.10. Boots fine with the “nouveau” graphics driver – graphics (even just menus) are very slow, choppy, and distorted. The three other driver options in Ubuntu (official NVIDIA drivers), all result in a variation of the black screen on boot up. There will be NO access to a command line/GUI in anyway what-so-ever (tried every option recommended out there, but the system is unusable at this stage). I can only reinstall, and try different drivers…and I only ever get one shot at it. QUESTIONS: -Does anyone know of a PROPRIETARY driver that will actually work on 12.10 with a NVIDIA or ATI card? -Should I just buy a newer graphics card to put in as a replacement? MORE INFO: This is my second computer, and I’m just trying to get a working install of Ubuntu on it. I don’t want to put much money into it, as I have seen Ubuntu run great on much older/less capable machines. I’ve got a decent'ish Core2Duo Intel processor (2.13Ghz), 2GB of RAM, 320GB hard drive, 32-bit architecture, and there is no other O/S installed. It appears as if the graphics card (NVIDIA Geforce 7350 LE) is holding me back. TRIED SO FAR: -all drivers available in Ubuntu *all fail -manual install of some different NVIDIA drivers *all fail -also tried installing the generic kernel, Nvidia driver doesn't work in 12.10 *no difference -tried installing 12.04 *same results -every method suggested to at least get a command line after switching to a NVIDIA driver *all fail -UPDATE- Re-tried everything above with a new NVIDIA Geforce 210...same results for everything. -UPDATE #2- Re-tried everything above with a new AMD Radeon HD 6450...installed the proprietary driver from Ubuntu's "Software Sources" menu...EVERYTHING NOW WORKS. See "answer" below for summary.

    Read the article

  • Making Thunderbird auto-add SMTP identities whenever I reply

    - by 0xC0000022L
    How can I teach Thunderbird to automatically add an SMTP identity whenever I reply to an email directed to <whatever>@<mydomain>? So if an SMTP is configured for <mydomain> but no identity exists for <whatever>@<mydomain>, how can I make Thunderbird dynamically recognize this and add it. Currently I have to manually add the identity every single time, but I would prefer it to to be added ad-hoc. As long as Thunderbird was configured to know about the SMTP serving <mydomain> this should be trivial, but I couldn't find an option. An add-on or something like a catch-all/wildcard identity would also do as long as it doesn't require manually setting up a new identity every time.

    Read the article

  • HAProxy overload protection

    - by user2050516
    using the HAProxy, would it be possible to configure an overload protection, to limit the amount of requests sent to the backing http server(s) to a given rate (z.B 100 Request per second ). If the threshold is exceeded requests should be answered with a default response. I am interested in requests per second not connections per second as a connection can have many requests. And yes to improve the servers is not an option here. If yes a configuration example to achieve that would be excellent. Thank you in advance.

    Read the article

  • Storing a pass-by-reference parameter as a pointer - Bad practice?

    - by Karl Nicoll
    I recently came across the following pattern in an API I've been forced to use: class SomeObject { public: // Constructor. SomeObject(bool copy = false); // Set a value. void SetValue(const ComplexType &value); private: bool m_copy; ComplexType *m_pComplexType; ComplexType m_complexType; }; // ------------------------------------------------------------ SomeObject::SomeObject(bool copy) : m_copy(copy) { } // ------------------------------------------------------------ void SomeObject::SetValue(const ComplexType &value) { if (m_copy) m_complexType.assign(value); else m_pComplexType = const_cast<ComplexType *>(&value); } The background behind this pattern is that it is used to hold data prior to it being encoded and sent to a TCP socket. The copy weirdness is designed to make the class SomeObject efficient by only holding a pointer to the object until it needs to be encoded, but also provide the option to copy values if the lifetime of the SomeObject exceeds the lifetime of a ComplexType. However, consider the following: SomeObject SomeFunction() { ComplexType complexTypeInstance(1); // Create an instance of ComplexType. SomeObject encodeHelper; encodeHelper.SetValue(complexTypeInstance); // Okay. return encodeHelper; // Uh oh! complexTypeInstance has been destroyed, and // now encoding will venture into the realm of undefined // behaviour! } I tripped over this because I used the default constructor, and this resulted in messages being encoded as blank (through a fluke of undefined behaviour). It took an absolute age to pinpoint the cause! Anyway, is this a standard pattern for something like this? Are there any advantages to doing it this way vs overloading the SetValue method to accept a pointer that I'm missing? Thanks!

    Read the article

  • How do I make control-click in Opera open new tabs?

    - by user21952-is-a-great-name
    Update: As of version 11.50, control-click opens the new tab in background, as desired. Yay! In Opera, as opposed to all other modern browsers I know, control-click does not open a new tab. Moreover, I couldn't find any good way to configure this behavior. The best available option seems to be this. However, it's so hacky that it won't work on HTTPS sites unless you enable user javascript there, which doesn't seem like a good idea. There also seem to be other proposed solutions, but none of them seem to work. I'd like this because my laptop has no middle mouse-button, and I'm a creature of habit. Do people have any ideas (for the latest version, 10.50)? I'd like ctrl-click to open a background tab, but I can do with foreground. Thank you!

    Read the article

  • How to meet Windows 8 upgrade's 20 GB requirement on a 40 GB SSD with a 22 GB Windows 7 install?

    - by deryus
    A PC I have has Windows 7 installed on a 40 GB SSD, and I bought a Windows 8 upgrade for it. The current Windows folder on it however is 22 GB, that's after removing hibernation, turning off the pagefile and removing all extra programs/features. So even if I purge every other file and folder, the Windows folder itself takes more than half the disk. The PC also has a 1 TB HDD, but the upgrade installer didn't give me any options about choosing another drive. So, is my only option to reinstall Windows 7 on a larger drive, then proceed with the Windows 8 upgrade? Or is there anything I can remove from the Windows folder that while might be dangerous for long term usage, is fine for the few minutes I need to get Windows 8 installing?

    Read the article

  • Changed folder icon in WinXP and now clicking a folder opens in a new window.

    - by widgisoft
    I got bored and decided to change my folder icon from the boring yellow default. Somewhere in my experiments, I discovered that anytime I open a folder it opens in a "search" window. This happened because when I go to the "(file folder)" type, the only option in there is "find" that is not supposed to be the default; upon saving this form XP tries to do me a favour and sets this as the new default. Now whenever I try to click a folder it opens the search window. How do I fix this?

    Read the article

  • Windows XP, have to use ctrl+alt+delete to log on as local administrator

    - by wickedj
    Hey, I have a weird issue, a user was was logging into a laptop using the local admin account which was working fine. I had to create another account on the system, which was also an admin account, when this happened the 'administrator' account disappeared from the 'choose an account to login with' screen. A quick workaround is available, if the user presses ctrl+alt+delete it brings you to the screen where you can type in the username and password, so by manually typing 'administrator' it can log in. Normally this would be easily fixed, I figured the admin account had somehow been disabled from the local system, but i checked all settings and it is setup fine. The laptop is not part of a domain, so I used the management console to delete the new account and all that succeeded in doing was making the 'choose an account to log in with' screen display no accounts to choose. So far I see nothing else to fix it, the option to change the default logon screen to style where you type the username and password also seems to be missing. any ideas?

    Read the article

  • Bringing my Dell XPS 13 ultrabook back to factory state

    - by TysHTTP
    I have a brand new Dell XPS 13 ultra book. After i picked it up at the store, i wiped the exising partitions which also contained the restore/rescue data, which you need to reinstall the factory version of Windows 8 that comes with this machine. I did this because we have a different MS partner edition of Windows which i prefer to run. After doing all this, i noticed that there was something wrong with machine. No real damage, but the specs are not completely as they should be. Turned out that a simple mistake while ordering. Now, long story short, the shop says that it has no problem with taking the product back, as long as it is in it's original state. And this is the problem i'm having, because i formatted the original partitions that contain that rescue option / windows 8 setup, i don't have a clue whether it's possible to get back to that original. Does anyone have an idea on how to get this fixed?

    Read the article

  • TPM had to be reintialized: Does a new recovery password have to be uploaded to AD?

    - by MDMoore313
    Some way some how, a user's machine couldn't get read the bitlocker password off of the TPM chip, and I had to enter the recovery key (stored in AD) to get in. No big deal, but once in the machine, I tried to suspend bitlocker per recovery documentation, and got an error message about the TPM not being initialized. I knew the TPM was on and activated in the BIOS, but Windows still made me reinitialize the TPM chip, and in the process it created a new TPM owner password. I found that odd because it prompted me to save this password or print it (there wasn't an option not to), but it made no reference of a recovery password, nor did it back this password up to AD. After the user took her laptop and left I started thinking that if the TPM password change, does the recovery password change also? If so, that new recovery password will need to be uploaded to AD, but MS' documentation doesn't make that clear, and doesn't back up the new recovery key (if one exists) to AD automatically when the group policy says it must, and from a network standpoint AD is accessible.

    Read the article

  • Unable to print A6 on Canon LBP 2900B php

    - by ????? ???
    I am working on web application. In which I have to print an invoice on A6 paper. To better control priniting I used TCPDF. In TCPDF options I have setup it to print A6. Now my TCPDF is working fine. But there is problem in printing it on Canon LBP2900B. As there is no option available to print A6 page. When I try to print on it. It used A4 paper to print it. I have also checked Custom options in Canon. But unable to figure out how to print it using Canon printer.

    Read the article

  • nslookup returning different results

    - by Sterno
    I'm having a problem with my machine. A site that I used to be able to access is no longer being found. After running nslookup, I notice that about 9 times out of 10 it returns a specific IP which times out if I ping it, but about 1 in 10 times it returns a different IP which works when I ping it. I've even tried adding an option parameter into nslookup to hit different DNS servers (such as Google's public DNS) and am getting the same result. Any idea what would cause the IP returned by DNS lookup to change back and forth like that?

    Read the article

  • Ubuntu cannot see Windows 7 partitions on install

    - by Nash0
    I've been trying to install Ubuntu 10.10 as a dual boot with Windows 7 on my Dell latitude e6510. It is currently running Windows 7 and I have used the MS disk tools to shrink the Win 7 NTFS partition to make room for Linux. The issue I'm having is that when I run Ubuntus installer by booting from CD it sees the entire hard drive as unallocated space. I have also tried Kbuntu 10.10, Fedora 14, booting a Gparted 0.8.0 usb drive, and Ubuntu "install in Windows" with wubi they all have problems. EDIT: When I run the "try Ubuntu" option on booting from cd it can mount my Windows partition and I can view the files. The output of sudo parted -l when running in try Ubuntu mode: Warning: /dev/sda contains GPT signatures, indicating that it has a GPT table. However, it does not have a valid fake msdos partition table, as it should. Perhaps it was corrupted -- possibly by a program that doesn't understand GPT partition tables. Or perhaps you deleted the GPT table, and are now using an msdos partition table. Is this a GPT partition table? Yes/No? yes Model: ATA ST9500420AS (scsi) Disk /dev/sda: 500GB Sector size (logical/physical): 512B/512B Partition Table: gpt Number Start End Size File system Name Flags Warning: Unable to open /dev/sr0 read-write (Read-only file system). /dev/sr0 has been opened read-only. Error: /dev/sr0: unrecognised disk label

    Read the article

  • HTTP PHP Authentication and Android

    - by edc598
    I am working on a website for which I hope to have an application for as well. Because of this, I am creating PHP API's which will go into my Database and serve specific data based on the method/function called. I want to protect these API's from misuse however, and I plan on implementing Authentication Digest to do so. However one of the OS's I want to support is Android. And I know that a malicious user would be able to reverse engineer the Android app and figure out my authentication scheme. I am left wondering: 1. Is there a better way to protect these API's from misuse? 2. Is there a way to prevent a malicious user from reverse engineering the app and potentially seeing the source code for it, enabling them to see my authentication scheme? 3. If none of these are preventable, then is my only option to have a Username/Password cred specifically for the Android app, and when eventually hacked, change the creds and issue an update for the app? I apologize if this is not the place to post such a question. Still pretty new to StackOverflow. Thanks in advance for any insight, it would be quite helpful.

    Read the article

  • Boot from Second SATA Drive

    - by Chris
    I have a Dell Precision 490 Workstation, and I just had my other question answered, Install Ubuntu to drive B without impacting drive A, and now I'm having a boot sequence issue. The external drive is great, boots up fine on my laptop, but how do I tell my desktop to boot from my second SATA drive and not the first SATAdrive. My drive configuration as follows SATA-0: Windows SATA-1: DVDR SATA-2: Ubuntu When I choose the boot menu, the option I have is "Internal Hard Drive". I assume it searches all drives, and loads the first bootable one it finds (which happens to be Windows), but I'd like to be able to select the drive from a list. Has anyone experienced this? Is possible without disabling the first hard drive in the BIOS?

    Read the article

  • Creating dynamic plots

    - by geoff92
    I'm completely new to web programming but I do have a programming background. I'd like to create a site that allows users to visit, enter some specifications into a form, submit the form, and then receive a graph. I have a few questions about this, only because I'm pretty ignorant: Is there a good framework I should start in? I know a lot of java, I'm okay with python, and I learned Ruby in the past. I figure I might use ruby on rails only because I hear of it so often and I think I've also heard it's easy. If anyone has some other recommendation, please suggest. The user will be entering data into a form. I'm guessing the request they'll be making should be one of a GET request, right? Because I don't intend for any of the data they're entering to modify my server (in fact, I don't intend on having a database). The data the user inputs will be used to perform calculations involving lots of matrices. I've written this functionality in python. If I use ruby on rails, should it be instead written in Ruby? Somewhere I've heard that you can either place the load of the work on your server or on the client's computer. Since the code performs heavy math, which option is preferable? How do I alter the setup to either make the client do the work or my server? Should I be using a "cgi-bin"? In the code that I have now, I use matplotlib, a python library, and then "show" the plot in order to see the graph. I specify the x and y limits, but I am able to "drag" the graph in order to see more data within the plot window. Ultimately, I want a graph to be shown on my site with the drag functionality. Is this possible? What if the client drags the graph so far, more computations must be made? Thanks!

    Read the article

  • Valid path to deploy vm on host

    - by ELSheepO
    I've recently added a VM to the library in SCVMM to use with MS Test Pro 2010, and run into a problem. I cannot import it using test pro, it just won't give me any option to import in the lab manager. Also, if I try to deploy it back to the host it came from, it gives me a error saying the path is not valid. Anyone have any insight into this? Also, SCVMM seems to freeze everytime I try to create a new VM from the template of the VM I've stored in the library, the same one thats giving me the problem when I try to deploy it on to the host. Thanks.

    Read the article

  • how to completly turn my mac into a host, mail server?

    - by idaho
    i have mac osx 10.6.3 (snow leopard) i have already setup mamp so it acts like a local server. but the next thing i wanna try doing is making it into a mail server where i can register an email locally maybe something like myname@myip or however its done. then having the ability to setup mx records, cname, and all that dns good stuff. this will only be for education purposes so that i can fully understand how it all works. what do i need to do? i read something about MailServeSnow but i cant get its domain to work , it must be down. would this be the best option? also would i need mac server or my os would be just fine? thanks

    Read the article

  • Visual Studio Development on Virtual Box, Boot Camp, or VMWare Fusion

    - by Eli
    I currently have a Mac, 2ghz and 2 gigs of ram, running OS X Leopard and Virtual Box with a Windows 7 Pro 32bit virtual machine. Performance on the virtual machine is fine for minor tasks but is very clunky while trying to multi-task or develop in Visual Studio 2008. What would be my best option for being able to use Visual Studio, keeping cost and time in mind? 1) Upgrade ram to 4 gigs ($100). Will this really improve my performance enough to use Visual Studio in a Windows 7 vm? Or am I just wasting time/money? 2) Reinstall/restore Windows 7 disk image as a Boot Camp partition. I assume this should improve my performance, yes? 3) Purchase VMWare fusion instead of VirtualBox. Does Fusion require less resources to run? I am open to any suggestions. Thanks in advance

    Read the article

  • Video/films organizing software for mac

    - by tig
    I am looking for application that will help me organize movies, clips and other videos I have. I tell it a folder to watch, and then I can mark every video there as watched/unwatched, set rating, add description, tag it. And possibility to open with preferred video player (for me — VLC). Option to get all info from imdb or other source will be very good. yFlicks would be a good one, but it uses quicktime, so it doesn't like any non standard codec and container (for example mkv), it doesn't work well on Snow Leopard. Any suggestions?

    Read the article

  • What are solutions and tradeoffs to maintain search result consistency in a web application

    - by iammichael
    Consider a web application with a custom search function that must display the results in a paged manner (twenty per page with up to hundreds of thousands of total results) and the ability to drill down to individual results that maintain next/previous links to navigate through the results. Re-executing the search on each page request to get the appropriate results for that page of data can be too expensive (up to 15s per search). Also, since the underlying data can change frequently (e.g. addition of new results), re-executing could cause the next/previous functionality to result in inconsistent behavior (e.g. the same results reappearing on a later page after having been viewed on an earlier page). What options exist to ensure the search results can be viewed across multiple pages in a consistent manner, and what tradeoffs does each option have in terms of network, CPU, memory, and storage requirements? EDIT: I thought caching the query search results was an obvious necessity. The question is really asking about where to cache the result set and what tradeoffs might exist to each. For example, storing the ids of the entities in the result set on the client, or storing the IDs of the entities themselves in the users session on the web server, or in a temporary table in the database. I'm not looking specifically for a single solution as different scenarios may result in different approaches (and such a question would be more suited for stackoverflow.com rather than here), but more of a design comparison between the possible approaches.

    Read the article

  • Solaris 11 VNC Server is "blurry" or "smeared"

    - by user12620111
    I've been annoyed by quality of the image that is displayed by my VNC viewer when I visit a Solaris 11 VNC server. How should I describe the image? Blurry? Grainy? Smeared? Low resolution? Compressed? Badly encoded? This is what I have gotten used to seeing on Solaris 11: This is not a problem for me when I view Solaris 10 VNC servers. I've finally taken the time to investigate, and the solution is simple. On the VNC client, don't allow "Tight" encoding. My VNC Viewer will negotiate to Tight encoding if it is available. When negotiating with the Solaris 10 VNC server, Tight is not a supported option, so the Solaris 10 server and my client will agree on ZRLE.  Now that I have disabled Tight encoding on my VNC client, the Solaris 11 VNC Servers looks much better: How should I describe the display when my VNC client is forced to negotiate to ZRLE encoding with the Solaris 11 VNC Server? Crisp? Clear? Higher resolution? Using a lossless compression algorithm? When I'm on a low bandwidth connection, I may re-enable Tight compression on my laptop. In the mean time, the ZRLE compression is sufficient for a coast-to-coast desktop, through the corporate firewall, encoded with VPN, through my ISP and onto my laptop. YMMV.

    Read the article

  • Migrate 3 terabytes of files to a new server windows 2003

    - by smackaysmith
    We have a new file server to handle the obscene amount of files generated by the company (PDFs, XLS, DOCs and JPGs). Files being moved to the new server total about 3tb. The problem is we can't take the company down for days to move the files. The other problem is the applications creating all these files have to reference previous files, so we can't simply point them to the new server. Also, there isn't an option to have the applications create files on the new server, but reference the old server for existing files. The servers are x64 win2003 r2. Both servers are on the same subnet. DFS doesn't work. Is there an application that can handle this amount of data to copy the files over, throttle bandwidth, and do a 'merge'? By merge I mean constantly copying over newly created files until the two servers are synched.

    Read the article

  • Somehow Google considers a properly 301'd URL as 200 and is still indexing the new content in old page?

    - by user2178914
    We redirected all the old URL's to new ones properly using htaccess. The problem is Google, somehow is still finding content in the old page(which it shouldn't) and stores it in the cache rather than the new URL. For eg: Old Page- http://www.natures-energies.com/iching.htm New Page- http://www.natures-energies.com/index.php?option=com_content&view=article&id=760 If you type the old URL into the browser it redirects If you fetch the old URL as Googlebot in the webmaster tools the header says 301/permanently redirected. If I try to crawl as any other bot it still says 301 redirected. Even if you click the old link in Google it redirects to the new URL. Only in its cache it shows the old URL and moreover it shows the new content in it! I am stumped on how Google manages to grab the new content and puts in the old URL instead of the new one! One more interesting thing is that if I try a cache for the new page it shows the cache of the new content with old URL! Any help would be appreciated. I am at end of my wits. I think i have tried almost everything. Is there anything that I'm missing to see? You can use this search to find the old url's. Maybe you'll some patterns that i missed. site:www.natures-energies.com inurl:htm -inurl:https|index

    Read the article

  • Why can't I install MySQL on my computer?

    - by Bea
    I have read a lot of tutorials, but I am still having problems. What I tried: I downloaded mysql-5.5.9-winx64. All that I read says that I can run Setup.exe, but there is no such file in the download. The other option I know there is, is including \mysql-5.5.9-winx64\bin in the PATH variable and then trying to execute the mysql command. When I do that, the error I get is: ERROR 2003 (HY000): Can't connect to MySQL server on 'localhost' (10061) I then downloaded mysql-5.5.9-winx64.msi, which is easier to install, but once I followed the instructions and it was installed, I got the same error executing the mysql command. How can I use MySQL? EDIT: I've now removed everything I installed, and I want to start from scratch.

    Read the article

< Previous Page | 433 434 435 436 437 438 439 440 441 442 443 444  | Next Page >