Search Results

Search found 13175 results on 527 pages for 'live backup'.

Page 165/527 | < Previous Page | 161 162 163 164 165 166 167 168 169 170 171 172  | Next Page >

  • Unable to connect to MSN network with Adium

    - by Adam Driscoll
    I'm new to the Mac world and tried both MSN Messenger for Mac (7.0 and 8.0 beta) as well as Adium to connect to the Windows network. I've enable 'Allow all incoming connections' in the Firewall settings. Windows Live Messenger works fine when connecting through the same router on my Windows laptop. I've triple checked my password and verified it through a web browser and Windows Live. Any ideas what my issue could be?

    Read the article

  • Using audio plugins in a rewire slave application

    - by Nik Reiman
    I would like to be able to use VST/AudioUnit effects from a rewire slave such as Ableton Live, but it seems that Live doesn't let you use plugins when running in this mode. I'm not exactly sure why that is, but it would be nice to get working so that I could use VST plugins from within GarageBand without having to use the VST2AU wrapper plugin. What sequencers let you use VST plugins when running as a rewire slave?

    Read the article

  • Performance of Cluster Shared Volume file copy from SAN

    - by Sequenzia
    I am hoping someone can help me out with a strange issue. We are running a Microsoft Failover Cluster with Server 2008 R2 and an Equallogic PS4000 SAN. Our main configuration has 2 Dell Poweredge T710 Servers in the cluster. We have CSV and Quorm setup. The servers each have 10 Broadcom 1Gb NICs. Right now 4 of the NICS are on the iSCSI network for accessing the SAN. They use MPIO and the Dell HIT pack. We have 5 VMs running on each node and everything runs smooth. No noticeable performance issues or anything. From the SAN I can see the 4 iSCSI connections from each server to each volume (CSV and Quorm). Again, it seems to perform great. The problem I am running into is with backups. I have tried a few backup programs like backupchain and Veeam. The problem is both of them are very very slow to backup the VMs. For instance I have a 500GB (fixed disc) VHD that’s running on the cluster. It takes over 18 hours to backup that VHD and that’s with compression and depuping turned off which is supposed to be the fasted. We also have a separate server that is just for backups. It has a lot of directed attached storage. As part of the troubleshooting I decided to bring that server into the cluster as a node. It now has access to the CSV and can read from C:\clusterstorage\volume1 which is where our VHDs live. This backup server only has 2 NICs. 1 NIC is going to the iSCSI network and the other is just on the main network. It has Intel NICS in it without any sort of MPIO or teaming. So with the 3rd server now in the cluster I started doing some benchmarking. I have a test VHD that’s about 7GBs that’s stored in the CSV. I have tested file copying that VHD from all 3 servers to directed attached storage in the respective server. The 2 Dell servers that are the main nodes in the cluster (they house the VMs) are reading that file at about 20Mbs/Sec. Which at that rate is way to slow for the backups. The other server which only has 1 NIC to the SAN is reading at around 100Mbs/Sec. I spent a few hours on the phone with Dell today about this . We went through all kind of tests and he was pretty dumb founded. He really has no idea why that server with only 1 NIC is reading about 5 times as fast as the servers with 4 NICS and MPIO. We looked at the network utilization of the NICs while the file copy was going on. The servers with the 4 NICs had a small increase of activity during the file copy but they only went up to around 8-10% on all 4 NICs. The other server with the 1 NIC jumped up to over 80% during the file copy. I plan on doing some more testing after hours and calling Dell back tomorrow but I really am confused (and so is Dell’s support rep) why I cannot get faster file copy access to the CSV on those servers. Anyone have any input on this? Any feedback would be greatly appreciated. Thanks in advance.

    Read the article

  • How to make a boot profile permanent?

    - by Usman Ajmal
    Hi, I have a Finnix live CD. I can customize it by remastering it. When I boot with the live CD I need to make a little change in the boot profile The boot profile before making the changes is linux apm=power-off vga=791 initrd=minirt quiet The boot profile after making the change become linux apm=power-off vga=791 initrd=minirt quiet root=/dev/sr0 Now, I need to make this change (adding root=/dev/sr0) permanent. How can I do that?

    Read the article

  • Ubuntu 10.04 Lucid on IBM Thinkpad T41

    - by naugtur
    I tried to boot my T41 from a Ubuntu Lucid liveCD and it worried me. The white text on violet background had green pixels around it. After the splashscreen went away Ubuntu popped an alert that the installer had some errors and will now run live. The live system worked fine I guess... Did anybody experience such behaviour?

    Read the article

  • Kernel-mode Authentication: 401 errors when accessing site from remote machines

    - by CJM
    I have several Classic ASP sites that use Integrated Windows Authentication and Kerberos delegation. They work OK on the live servers (recently moved to a Server 2008/IIS7 servers), but do not work fully on my development PC or my development server. The IIS on both machines were configured through an IIS web deployment tool package which was exported from an old machine; the deployment didn't work perfectly, and I had to tinker a bit to get the sites working. When accessing the apps locally on either machine, they work fine; when accessing from another machine, the user is prompted by a username/password dialog, and regardless of what you enter, ultimately it results in a 401 (Unauthorised) error. I've tried comparing the configuration of these machines against similar live servers (that all work fine), and they seem generally comparable (given that none of the live servers are yet on IIS7.5 (Windows 7/Server 2008 R2). These applications run in a common application pool which uses a special domain user as it's identity - this user has similar permissions on the live and development machines. On IIS6 platforms, to enable kerberos delegation, I needed to set up some SPNs for this user, and they are still in place (even though I don't believe they are needed any longer for IIS7+ due to kernel-mode authentication), Furthermore, this account is enabled for Kerberos delegation in Active Directory, as is each machine I am dealing with. I'm considering the possibility that the deployment might have made changes/failed to make changes to the IIS configuration thus causing this problem. Perhaps a complete rebuild (minus another web deployment attempt) would solve the problem, but I'd rather fix (thus understand) the current problem. Any ideas so far? I've just had another attempt at fixing this issue, and I've made some progress, but I don't have a complete fix...yet. I've discovered that if I access the sites via IP address (than via NetBIOS name), I get the same dialog, except that it accepts my credentials and thus the application works - not quite a fix, but a useful step. More interestingly, I discovered that if I disable Kernel-mode authentication (in IIS Manager Website Authentication Advanced Settings), the applications work perfectly. My foggy understanding is that this is effectively working in the pre-IIS7 way. A reasonable short-term solution, but consider the following explicit advice from IIS on this issue: By default, IIS enables kernel-mode authentication, which may improve authentication performance and prevent authentication problems with application pools configured to use a custom identity. As a best practice, do not disable this setting if Kerberos authentication is used in your environment and the application pool is configured to use a custom identity. Clearly, this is not the way my applications should be working. So what is the issue?

    Read the article

  • Windows Media Center Buffering and Asus MyCinema U3100mini

    - by dsimcha
    I'm running an Asus MyCinema U3100mini ATSC on Windows 7 64-bit. When I play live TV in Windows Media Center, it's very choppy and uses 500+ MB of RAM, I'm guessing due to the hard drive buffering functionality. Is there any way to disable the live TV pause buffer completely? If not, can anyone recommend alternative software that: Works with the MyCinema. Is lightweight and not horribly bloated with features I'll never use like Windows Media Center is.

    Read the article

  • How can I capture a video with Windows 7

    - by user12029
    I am trying to capture a live video from my digital camera using Windows 7. Live Movie maker and Windows Movie maker 2.6 do not have "capture" menu item. Apparently, Windows Movie maker 2.1 has a "capture" menu item. However, I don't know how to download it on Windows 7. Any suggestion?

    Read the article

  • Secure linux distro that sends traffic through Tor

    - by gravestone
    I'm looking for a live Linux distro that is secure and preserves my anonymity online. For what I need, Tor seems to do the job – but such software requires configuration that someone who is not knowledgeable in how it works might find difficult. I was looking at 'Lightweight portable security' but it does not specify whether it sends all traffic through Tor. Does anyone know more information about it and/or any other such live Linux distro?

    Read the article

  • Creating Test Sites

    - by Robert
    I have a website running off site. When we hire someone I would like to create a test site (a copy of live site) for the new employee to tinker with. I will need to take fresh copies of the Files and Database (basically a snapshot) and allow them to access these copied files and database so they could edit and upload them to see the changes they made as if it was the live site Basically what is the best practice for creating a copy of a website for testing? Server is running Linux, PHP, mySQL

    Read the article

  • SQL Server Log File Won't Shrink due cause "log are pending replication" on non replicated DB?

    - by user796466
    I have a non Mission Critial DB 9am-5pm SQL Server database that I have set up to do nightly full backups and log backups every 30 minutes during business hours. The database is in full recovery and normally I have no reason to truncate/shrink logs unless I do some heavy maintenance. Log backups manage the size with no issue. However I have not been at this client for several weeks and upon inspection I noticed that the log had grown to about 10 times the size of the .mdf file. I poked around backups had been running and I had not gotten any severity error alerts (SQL mail). I attempted to put DB in simple recovery and shrink the log, this was no good. I precede to try a log backup and I got: The log was not truncated because records at the beginning of the log are pending replication or Change Data Capture. Ensure the Log Reader Agent or capture job is running or use sp_repldone to mark transactions as distributed or captured. Restart SQL Server rinse repeat same thing ... I said ??? Replication is not nor ever has been set up on this DB or database /server ??? So the log backups have not been flushing the .ldf. So I did a couple hours of research and I found: http://www.sqlmonster.com/Uwe/Forum.aspx/sql-server/5445/Log-file-is-not-truncated-inspite-of-regular-log-backup http://www.eggheadcafe.com/software/aspnet/30708322/the-log-was-not-truncated-because-records-at-the-beginning-of-the-log-are-pending-replication.aspx seems to be some kind of poorly documented bug ?? The solution seems to have been to run exec sp_repldone, more precisley EXEC sp_repldone @xactid = NULL, @xact_segno = NULL, @numtrans = 0, @time= 0, @reset = 1 This procedure can be used in emergency situations to allow truncation of the transaction log when transactions pending replication are present. Using this procedure prevents Microsoft SQL Server 2000 from replicating the database until the database is unpublished and republished. ~ MSDN When I do that I get the following Msg 18757, Level 16, State 1, Procedure sp_repldone, Line 1 Unable to execute procedure. The database is not published. Execute the procedure in a database that is published for replication. Which makes sense Because the DB has never been published for replication. I have several questions: A) First and foremost is, WTF is going on ? What is causeing this, I am interested in knowing the why here ? Is this genuinley a bug or is there some aspect of the backup that is not functioning properly that cause's the DB to mimick a replicated state ? Someone please edify me on this. B) Second ... Do I really have to publish / replicate this DB to exec this SP to fix this ??? Sounds crazy or is there some T-SQL that I can put it in a published state exec the proc and be on my way ... C) Third, if I do indeed have to publish this database to exec the SP to release this unneeded mis replicated/intended log , to get my .ldf file and backup back on track. How do I publish the database without an online host that it is asking for ??? I don't generally do this kind of database administration and need some guidance. Sorry if this is too verbose but just voicing the question helps me clarify it ... Thank you in advance for your help

    Read the article

  • Is it possible to change the Raid5 chunk size of an existing device?

    - by AlexCombas
    I have an existing raid5 device which I created using mdadm on Linux. When I created the device I set the chunk size to 64 but I would like to test the performance of various sizes but I don't want to have to rebuild my entire system to do so. If it is not possible to do it live then is it possible to do this by booting with a rescue disk? Any advice on the steps how to do this, either live or with a rescue disk, will be greatly appreciated.

    Read the article

  • How to format pendrive from fat32 to ext3 in windows7

    - by newb
    I am trying to make a live usb of OPHCRACK and tried to boot from FAT32 pendrive. But after making live usb and boot from it the ophcrack didnt work. After searching a while i came to understand that ophcrack will not work in a fat32 pendrive and we have to convert it into ext3. But i am getting hard time finding a method or software which can be used to convert fat32 pendrive to ext3 in windows 7. Can you suggest any method or software's for this purpose

    Read the article

  • Adding additional locations to Office 2013 save "Places"

    - by Paperjam
    When saving a document in Office, you are given the option of saving locally or to your SkyDrive account. Presumably, it's possible to add additional locations to this menu (Dropbox, etc.). Is this possible in the Live Preview, and how does one accomplish this? Edit: when I asked this question, I was not logged into a Live account. I have since logged in and discovered services may be added to the logged-in account, however Dropbox (and non-MS services) are not offered. Account settings:

    Read the article

  • Do I need the meta-key for vim?

    - by Riyaah
    I'm looking in to learning emacs or vim. I started out with emacs but found the need for a meta key to be a hassle, especially since I have a non-english keyboard layout on my macbook. So far I haven't seen any references to meta in vim, so my question is: Can I live without meta in vim? If so that'll settle the vim vs. emacs question for me, otherwise I'll just have to learn to live with some workaround.

    Read the article

  • Overwriting the content from one MOSS content database to another

    - by 78lro
    We have a content database on our live moss server. It contains one site collection with several sub-sites. I'm using the stsadm export command to produce a cmp file, then moving this to our test server in a different farm. I then want to import this content into the content database on our test farm, using the import stsadm command results in me being left with all the existing test data as well as the live data. I tried detaching the existing content database from test in central admin and creating a new empty one,to the then run the import against that but the import failed as obviously there's not root site in the empty db. The aim is to have the data on test look like live, clearing out all the test data. Can anyone suggest a good approach to this type of problem?

    Read the article

  • HttpWebRequest to different IP than the domain resolves to

    - by fyjham
    Hey, Long story short an API I'm calling's different environments (dev/staging/uat/live) is set up by putting a host-record on the server so the live domain resolves to their other server in for the HTTP request. The problem is that they've done this with so many different environments that we don't have enough servers to use the server-wide host files for it anymore (We've got some environments running off the same servers - luckily not dev and live though :P). I'm wondering if there's a way to make WebRequest request to a domain but explicitly specify the IP of the server it should connect to? Or is there any way of doing this short of going all the way down to socket connections (Which I'd really prefer not to waste time/create bugs by trying to re-implementing the HTTP protocol). PS: I've tried and we can't just get a new sub-domain for each environment.

    Read the article

  • get img src from an img loaded dynamically using jQuery

    - by Lee
    Hey Guys (and Ladies) I am looking at getting the image name from an image with a src loaded dynamically. Basically what I am doing is using the Google chat badge on a site to live chatting. And else where on the page, I have an image saying Live Chat:Online or Live Chat:Offline. And I want this to change depending on whether I am available to chat or not. Does this make sense? Anyway, the easiest way I figure to do this would be check the img url. If the img is offline.gif then obviously I am offline. An example of a dynamically loaded img would be something like <img src="http://www.google.com/talk/service/badge/Show?encrypted_acount_id_here" /> Once this image has loaded, it loads one of the following images "http://www.google.com/talk/service/resources/offlinef.gif" "http://www.google.com/talk/service/resources/idlef.gif" "http://www.google.com/talk/service/resources/onlinef.gif" Hopefully this make sense now. Thanks heaps

    Read the article

  • Finding App Id under .ipa or .app

    - by Rafael Oliveira
    I'm building an application that search and recognizes any iPhone apps that user has in his/her computer. I would like to know a way to "extract" the id of the application from the .ipa file. I was trying to do the recognition using only the App File Name, but I discovered that the File Name is not the name of the app in Apple Store. Live Poker 6K Free by Zynga != Live Poker 3.7.ipa The id i'm talking about is the app id, like in, http://itunes.apple.com/app/live-poker-6k-free-by-zynga/id354901953?mt=8 the id is 354901953. Does any body has a clue how can I manage to find this information?

    Read the article

  • Does Safari have a timeout issue?

    - by woodysapsucker
    I have a website that works fine in Firefox and IE but never finishes loading in Safari 4.0.4. The live website loads a menu then loads a Google map. Using Safari, the Google map never loads. I have a test website on the same server that uses the exact same code for loading a Google map and Safari can load the map (this is the one that won't load on my live web). This test website does not have the menu - only the Google map. In both cases (live and test) the main web page calls a loadmap.js file. I've been pulling everything apart to try to identify why Safari won't finish loading. Has anyone run across any problems/solutions with Safari that may help me find a resolution to this problem?

    Read the article

  • Searching a key word with java client programme from wsdl

    - by samy
    Heyy I have a problem I generated a client programme using http://api.search.live.net/search.wsdl this service for searching a key words.. I generated client by help of Eclipse-web project from this service. I have done searched on this service(live.net) but i can't show on the console. How i can do that? public static void main(String[] args) throws RemoteException { LiveSearchPortTypeProxy bb=new LiveSearchPortTypeProxy(); SearchRequest request=new SearchRequest(); SearchRequestType1 bbs=new SearchRequestType1(); aas.setParameters(request); sorgu.setAppId("*****************************************"); //you can take this ID from live service for using this service sorgu.setSources(new SourceType[]{SourceType.Web}); sorgu.setQuery("keyword"); SearchResponseType0 cevap= bb.search(bbs); }

    Read the article

  • Regarding Multiple Tab-Class

    - by Christine
    Hi! I've problem regarding GUI with one Menu and one Order Class. I've created a variable to store how many items have been selected in the Menu Class. private int totalSelected; The var totalSelected is live updated. It can be changed anytime depending on actionPerformed() function.(Exp: totalSelected will add up all the selected items) In the Order Class, how can I access to the live update variable totalSelected in order to retrieve the live update value? When I invoke getTotalSelected() function inside the Menu Class, I will only obtain a 0 value. Thanks for your help ^^!

    Read the article

< Previous Page | 161 162 163 164 165 166 167 168 169 170 171 172  | Next Page >