Search Results

Search found 21331 results on 854 pages for 'require once'.

Page 463/854 | < Previous Page | 459 460 461 462 463 464 465 466 467 468 469 470  | Next Page >

  • How can I prevent JungleDisk/MacOS X (10.6) creating a local volume for a removed external drive?

    - by Rew
    Ok, here is situation: I use JungleDisk to sync an online folder on to a external drive connected to my Mac. If I right click Finder, click Go to Folder... then type /Volumes/ I see the drive linked here. Once I remove the external drive, an actual folder is created here in the name of the external drive, JungleDisk continues to copy files to this folder, rather than stop. Is this a feature of Mac OS X? Can I turn if off? After I re-connect my external drive, the link to the drive is appended with a 1 (so if I called the drive SpareDrive it becomes SpareDrive 1 as the newly created folder is called SpareDrive. I realise my explanation isn't very clear, but anyone understand this, and knows how to prevent it happening please let me know. PS: I have a low reputation as I don't use this often, I tend to use stackoverflow, but will check back here for answers.

    Read the article

  • How can I search for Gmail conversations that have more than one e-mail? [closed]

    - by Matt Refghi
    I previously used a service that sent me Slashdot articles as individual e-mails. All I had to do was give them the RSS url, and they handled the rest. On the Gmail side, I simply made a filter, labeling these e-mails as "News". Once I stopped using the service, I realized that I had more than 20,000 such e-mails stockpiled. While I could delete all "News" e-mails, there are some that I've forwarded to friends, sometimes receiving multiple replies. These conversations are valuable - I would like to keep them. As far as I can tell, there's no way to search for conversations that have more than one e-mail in the thread. Up until now, I've been deleting them page by page, carefully de-selecting the conversations that have more than one e-mail; however, it will take far too much time to complete. Is there an easier way that I'm missing? Thanks.

    Read the article

  • User sees all Sharepoint document as being from 1938

    - by Ulys
    Hello, I currently have a problem I was unable to solve concerning Sharepoint. One of our user is seeing every file on Sharepoint as being from around 1938. The Created and Last Modified field are affected. When the user saves a file on his computer, the date is correct. If I log on his computer, the date are fine. If he log on my computer, the date are wrong once again, so it is linked to its Sharepoint account. The day and month are wrong to, but files uploaded 3 days apart are still 3 days apart for him. I don't see how anything could affect a date like that... Any help would be appreciated. Thanks, PS: Destroying his account and give him a new one is not a solution for now. He has been in this company for a long time and he has a lot of rights inherited from Sharepoint 2003 that are complicated to create.

    Read the article

  • Possible causes for Domain server being unavailable?

    - by serversurfer
    One of our servers was compromised after a user with administrative privileges accidentally loaded a virus from a USB drive on a desktop connected to the domain. The two most obvious symptoms of this were: The server is no longer responding to login attempts The root directory of the drive containing user data has been filled with randomly named empty folders. (Initially it was around a million folders, I've been slowly deleting them.) I've run several virus scans from different vendors and am fairly confident the virus has been removed but the damage is done. I'm hoping the two symptoms are related and that once the directories are gone the server will start responding again. The drive is very slow to respond. I'm deleting about 20k folders at a time. Anymore than that and windows explorer becomes unresponsive. In the event that I finish cleaning up the HD and things don't return to normal what other things can I check?

    Read the article

  • Do there exist programming languages where a variable can truly know its own name?

    - by Job
    In PHP and Python one can iterate over the local variables and, if there is only once choice where the value matches, you could say that you know what the variable's name is, but this does not always work. Machine code does not have variable names. C compiles to assembly and does not have any native reflection capabilities, so it would not know it's name. (Edit: per Anton's answer the pre-processor can know the variable's name). Do there exist programming languages where a variable would know it's name? It gets tricky if you do something like b = a and b does not become a copy of a but a reference to the same place. EDIT: Why in the world would you want this? I can think of one example: error checking that can survive automatic refactoring. Consider this C# snippet: private void CheckEnumStr(string paramName, string paramValue) { if (paramName != "pony" && paramName != "horse") { string exceptionMessage = String.Format( "Unexpected value '{0}' of the parameter named '{1}'.", paramValue, paramName); throw new ArgumentException(exceptionMessage); } } ... CheckEnumStr("a", a); // Var 'a' does not know its name - this will not survive naive auto-refactoring There are other libraries provided by Microsoft and others that allow to check for errors (sorry the names have escaped me). I have seen one library which with the help of closures/lambdas can accomplish error checking that can survive refactoring, but it does not feel idiomatic. This would be one reason why I might want a language where a variable knows its name.

    Read the article

  • XPS M1730 Freezing & Crashing unexpectedly

    - by evesirim
    I have a Dell XPS M1730, freshly installed with Windows 7 Professional 32-bit that every so often crashes completely unexpectedly. It need not be running any applications at all for this to happen, never mind anything particularly intensive. It also clearly is not a BSoD issue as no Minidump files are being created in the process. All drivers are up to date and have been both rolled back and then updated and the Device Manager reports no issues. There are very few applications installed on the machine and it has only been online once. It runs a minimal boot of Win7 and yet still after all of this something is making it crash. All of this suggests to me it may be more like a hardware problem but to be honest I really have no idea - so I thought I'd try here. Any thoughts...? Many thanks in advance

    Read the article

  • using apple-mobile-web-app-capable and cache.manifest issue [migrated]

    - by LocoMike
    So I have this simple html file <!DOCTYPE HTML> <html manifest="cache.manifest"><head> <meta name="apple-mobile-web-app-capable" content="yes"> <meta name="apple-mobile-web-app-status-bar-style" content="black"> <title>Test</title> <meta http-equiv="content-type" content="text/html"> <meta name="HandheldFriendly" content="true"> <meta name="viewport" content="width=320; initial-scale=1.0; maximum-scale=1.0; user-scalable=0;"> <style type="text/css"></style></head> <body marginwidth="0" marginheight="0" topmargin="0" leftmargin="0"> <h1>hello</h1> </body> </html> My cache.manifest is simply CACHE MANIFEST I run this website on my local server (localhost). I load it from iphone safari and it works fine. I then stop the server and load it again, and it works, because the offline cache is doing its job. However... if I save the website as a start icon in the iphone dashboard, and then I try to open it with the server stopped it won't load. However... if I open it with the server running at least once (it will work) then I can open it later without problem. It looks like even though the page was cached in safari, it is not cached in this saved app. Anybody knows how to get around this? Thank you!

    Read the article

  • USB drivers stopped working in Windows 8

    - by Maxim V. Pavlov
    I have installed an MSDN version of Windows 8 Professional (x64) RTM. For about 3 restarts everything worked well. But once I've rebooted again - USB drivers (all of them, inluding the USB 3 drivers) stopped working. Device Manager properties said that the USB driver is not compatible with the system. Gigabyte doesn't list Windows 8 drivers for my MB X58A-UD3R. Tried to reinstall windows 7 USB drivers - system said the current driver is fine and didn't want to reinstall. Is this the common Windows 8 problem? How can solve it or debug it even more?

    Read the article

  • Is there a way to schedule an automatic WinClone run on my Bootcamp partition?

    - by user17873
    Hi, I've currently got time machine setup to back up my entire OS X installation. I also have a backup tool within my Bootcamp Windows 7 installation which automatically backs up my windows profile data to an external drive partition. Finally, I'm also backing up my Bootcamp partition weekly and storing on an external drive using WinClone. The final piece I need to complete my external backup process is to have the WinClone application backup my Bootcamp partition automatically once a week rather than having to call it manually and remember. Is this possible?

    Read the article

  • MySQL Replication fix after server shutdown/start

    - by Jagbir
    Server1 is Master Server2 is Slave Both are in our AWS testing env and we stop them once done with our work. When start again, Master rotates/creates new binary log file but slave keep looking for same/existing one and replication stops. Right now, I'm manually repairing it by (slave): stop slave; CHANGE MASTER TO MASTER_HOST='xx', MASTER_USER='xxx', MASTER_PASSWORD='xxx' , MASTER_LOG_FILE='new-mysql-bin.00000x',MASTER_LOG_POS=107; start slave; show slave status\G and slave becomes good again.Mysql is 5.5.x on Ubuntu 12.04. Will appreciate any help in automating it.

    Read the article

  • TPM had to be reintialized: Does a new recovery password have to be uploaded to AD?

    - by MDMoore313
    Some way some how, a user's machine couldn't get read the bitlocker password off of the TPM chip, and I had to enter the recovery key (stored in AD) to get in. No big deal, but once in the machine, I tried to suspend bitlocker per recovery documentation, and got an error message about the TPM not being initialized. I knew the TPM was on and activated in the BIOS, but Windows still made me reinitialize the TPM chip, and in the process it created a new TPM owner password. I found that odd because it prompted me to save this password or print it (there wasn't an option not to), but it made no reference of a recovery password, nor did it back this password up to AD. After the user took her laptop and left I started thinking that if the TPM password change, does the recovery password change also? If so, that new recovery password will need to be uploaded to AD, but MS' documentation doesn't make that clear, and doesn't back up the new recovery key (if one exists) to AD automatically when the group policy says it must, and from a network standpoint AD is accessible.

    Read the article

  • What's faster, cp -R or unpacking tar.gz files?

    - by Buttle Butkus
    I have some tar.gz files that total many gigabytes on a CentOS system. Most of the tar.gz files are actually pretty small, but the ones with images are large. One is 7.7G, another is about 4G, and a couple around 1G. I have unpacked the files once already and now I want a second copy of all those files. I assumed that copying the unpacked files would be faster than re-unpacking them. But I started running cp -R about 10 minutes ago and so far less than 500M is copied. I feel certain that the unpacking process was faster. Am I right? And if so, why? It doesn't seem to make sense that unpacking would be faster than simply duplicating existing structures.

    Read the article

  • Amazon EC2 Sign In

    - by Barry
    When I change the home directory of my Amazon EC2 instance from /home/ubuntu to /home/ubuntu/folder in the /etc/passwd file, I am no longer able to access the instance using my existing keypair. Once I switch it back to the original directory I have no problems and can log into my instance as normal. I have checked the permissions on the new folder and they are drwxr-xr-x, which is the same the /home/ubuntu folder. I have a number of instances running at the minute and because of this change I have no way of logging back into them to rectify the situation. Does anyone have an idea what is going on? Thanks in advance

    Read the article

  • KVM Switch for a mac and a Windows (XP or 7) machine to share VDU

    - by Adrian Parker
    Have a MacPro (snow leopard) connected to an (windows standard) Asus 25" monitor via a DVI--VGA adapter. Now the boss wants me to work from home, so I want to share my Asus display with a Windows XP laptop. No doubt once my wife sees this, she will want to do the same thing, but with a Windows 7 laptop. So what I would like, I think, is your recommendations for a KVM switch (or better solution) that allows the Mac and a (windows 7 or windows xp) laptop to share the Asus display. Bonus marks if they can share Apple keyboard and magic mouse, but am quite happy to use separate mouse keyboards. The MacPro is the one that is always connected, the laptops come and go. Any ideas? Thanks in advance for any help.

    Read the article

  • Installing drivers and getting -> Error Code 28

    - by Adrov
    I've recently upgraded mainboard, without reinstalling OS, so I guess that's the issue. I really don't want to install OS at this moment. Issue is I can't install USB drivers, if I right-click uninstall, just installs same driver, which isn't working ofc. and giving error code 28. I fixed such issue once long long time ago, with editing registry, but I really can't remember what and where I have to do, so if anyone know please let me know, I'm also open to all other solutions to this issue. http://i.stack.imgur.com/AuBtB.png

    Read the article

  • Can I share data between two users in an ASP.NET Application?

    - by Dave
    I have an issue with Roles.IsUserInRole function. It take hell amount of time to just check if the logged-in user is in particular role(typ 3-9 sec). I searched to find a solution and arrived on this but I am not sure If I have fully grasped it. What I got from the above, A new derived class is created. Inside that class, there is a list which retrieves all user at once. The next time you check IsUserInRole, you do not use the actual IsUserInRole method but rather use the one you overrode in your class. Is this the correct description? Am I on track? My question is, can data be share between two different users in ASP.NET application? If yes, will the shared data exist only if there is at least one user logged in. If all users logs out, that shared data is destroyed? My point is this data will be created only one time whenever a user logs in. For all subsequent users they can use this data and check their roles against the list? I need a detailed answer. My application has users and different roles. We are using ASP.NET roles.

    Read the article

  • Getting "is not a valid mailbox" when email sent to primary address

    - by Juan Tarquino
    Anyone outside my domain who sends me an email is getting a bounced email with the following error from Exchange: 550 550 5.1.1 [email protected] is not a valid mailbox (state 14). I can't receive emails from external addresses until I send an email to one of my secondary addresses. Once I sent it to my secondary address, my primary address will start working for a while and fail again after a few hours. We use Exchange 2003. Anyone has any suggestions on what to try?

    Read the article

  • Upgrading a hard disk – To repave or to migrate, that is the question

    - by guybarrette
    I recently changed my laptop hard disk from the stock 250GB 5400 drive to a 320GB 7200 drive.  And no, I didn’t bought a SSD drive because the cost is way too much right now.  At $70, my upgrade was a lot cheaper than a SSD drive.  Maybe next year. When changing a system main hard drive, one must ask himself: To repave or to migrate, that is the question.  I choose to migrate so I went to the Acronis Website to take a look at their product line.  They have a few products that could do the job.  One being Acronis Migrate Easy 7.0 and the other being Acronis True Image Home 2010.  Since True Image was just $10 more then Migrate Easy, I bought True Image. I inserted my new hard drive in a 2.5” USB enclosure, and started the migration process.  Once the data copied, I switched the drives.  The process went very smoothly and without hiccups.  Highly recommended. BTW, Acronis offers free trials so I guess that nothing can stop you from “testing” a migration  ;-) var addthis_pub="guybarrette";

    Read the article

  • Why can't I install MySQL on my computer?

    - by Bea
    I have read a lot of tutorials, but I am still having problems. What I tried: I downloaded mysql-5.5.9-winx64. All that I read says that I can run Setup.exe, but there is no such file in the download. The other option I know there is, is including \mysql-5.5.9-winx64\bin in the PATH variable and then trying to execute the mysql command. When I do that, the error I get is: ERROR 2003 (HY000): Can't connect to MySQL server on 'localhost' (10061) I then downloaded mysql-5.5.9-winx64.msi, which is easier to install, but once I followed the instructions and it was installed, I got the same error executing the mysql command. How can I use MySQL? EDIT: I've now removed everything I installed, and I want to start from scratch.

    Read the article

  • nginx block URI request but allow internal directory

    - by Mike Anders
    I'm new to nginx from apache. I'm trying to simply block the URIs: /_mydir/* = / (redirect) But, I want to rewrite: /ex/(.*)$ = /_mydir/$1 I have tried: location /ex/ { rewrite ^/ex/(.*)$ /_mydir/$1 last; } location /_mydir { rewrite ^/_mydir/(.*)$ http://$http_host/ redirect; } But what always happens is once I block the '/_mydir' directory the rewrite is also blocked. I have also tried: location /_mydir/ { internal; } This also ends up blocking the rewrite. All help is greatly appreciated, thanks. UPDATE: I fixed this problem using: rewrite ^/ex/(.*)$ /_mydir/$1 break;

    Read the article

  • Help with Abstract Factory Pattern

    - by brazc0re
    I need help with a abstract factory pattern design. This question is a continuation of: Design help with parallel process I am really confused where I should be initializing all of the settings for each type of medium (ex: RS232, TCP/IP, etc). Attached is the drawing on how I am setting up the pattern: As shown, when a medium is created, each medium imposes a ICreateMedium interface. I would assume that the Create() method also create the proper object, such as SerialPort serialPort = new SerialPort("COM1", baud); however, TCPIPMedium would have an issue with the interface because it wouldn't need to initialize a serial port object. I know I am doing something majorly wrong here. I just can't figure it out and have been stuck for a while. What I also get confused on show the interface IMedium will get access to the communication object once it is created so it can write out the appropriate byte[] packet. Any guidance would be greatly appreciated. My main goal is to have the Communicator class spit a packet out without caring which type of medium is active.

    Read the article

  • Windows 7 Explorer - Activate mouse focus to be able to scroll in tree and file view?

    - by none
    I'd like to be able to scroll in the tree view without having to click in it. Is there a way to do this? I have once used a tool that generally gives focus to the window underneath the mouse cursor, but this caused some other glitches so I would like to achieve this without an extra tool. I also think that there are programs that embed the Windows Explorer and offer more features, including the behaviour I would like to have. But maybe a registry value change is all that is needed?

    Read the article

  • MySql transfer / update (a bit specific)

    - by Jeff
    before posting I was digging whole site but didn't find help for my problem, so I hope someone will help... Facts: 30 Gb mysql database on remote server (about 20.000.000 rows) data are once weekly updated in local network (mysql) I need to transfer/replace local updated database with remote connection is about 2mb (real mb, not mbps) up/down Point is that I can't have 'down time' of remote mysql server. Until now I Tried: navicat data sync - Ok, but take about 3 days to finish dbForge - ok but need 5 days to finish mysql dump transfer to remote server and execution - about day, but a lot of downtime rsync folder with database /mysql/lib/MY_DATABASE - 4 hours, but after that I need to execute always 'repir on remote server' which takes about 2 hours, and a lot of down time mysql dump piped from cl to directly goto server - still now satisfied many problems I could give you more things that I tried... mysql replication - slow Anyase, what is best,best way to: refresh remote mysql on weekly level and in same time to have 0 sec down time nor huge server load If you have any idea please share

    Read the article

  • Apache has no rights to unzip file uploaded by DirectAdmin created FTP user

    - by FlyOn
    I know similar things have been asked a thousandfold, but I can't seem to figure it out.. I have recently started renting a VPS. It runs DirectAdmin. I created a reseller account in the admin account, and created a new user (with FTP account) with the reseller account. I logged in with FTP as the created user and uploaded a zip file and a php file which will extract the zip. But when running the script in the browser I get "Permission denied" errors however. Probably because Apache is not in the same group as the users FTP account. Now my question is: How can I solve this once and for all, so that I don't have to do it again for every new user account I create through DirectAdmin?

    Read the article

  • Postfix - How to process incoming emails?

    - by Borivojevic
    Hello Does anybody know how to process incoming emails for virtual mailboxes in postfix? I am building web application where users add new content by sending emails to application. Email address used for each user is custom (eg. [email protected]) and it is dynamically created as a Postfix virtual mailbox. User needs to be able to send email to his custom mailbox address ([email protected]) and i want to process each incoming email, parse it's contents and populate my database with data from email. I tried using Postfix After Queue filter but what i really wont is to process emails once they are saved in users virtual mailbox folder.

    Read the article

< Previous Page | 459 460 461 462 463 464 465 466 467 468 469 470  | Next Page >