Search Results

Search found 335 results on 14 pages for 'mirroring'.

Page 5/14 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • User mapping lost after manual failover

    - by fordan
    I have two Microsoft SQL Server instances set up for mirroring each with a number of databases. There are a number of logins and for each database one or more user/login mappings. When I restore a backup of database I always have to redo the login/user mappings. I understand this because the logins are per database server. So after restoring the databases on the pricipal I redid the login/user mappings. This was not possible for the mirror because the databases were 'restoring'. After a manual failover I could not use the databases because user credentials were missing. This was not unexpected, so I did the login/user mapping again. I did a manual failover again to make the initial pricipal, which was now the mirror, principal again. To my surprise I could not use the databases because the login/user mappings were gone. Is this the expected behaviour?

    Read the article

  • SQL Server Database In Single User Mode after Failover

    - by jlichauc
    Here is a weird situation we experienced with a SQL Server 2008 Database Mirroring Failover. We have a pair of mirrored databases running in high-availability mode and both the principal and mirror showed as synchronized. As part of some maintenance I triggered a manual failover of the principal to the mirror. However after the failover the principal was now in single-user mode instead of the expected "Principal/Synchronized" state we usually get. The database had been in multi-user mode on the previous principal before this had happened. We ended up stopping all applications, restarting the SQL Server instances, and executing "ALTER DATABASE ... SET MULTI_USER" to bring the database back to the expected "Principal/Synchronized" state in a multi-user mode. Question. Does anyone know where SQL Server stores information about whether a database should be in single-user mode or not? I'm wondering if there is some system database or table that has this setting recorded somewhere. In particular we had an incident once with the database on the original principal (the one I was failing over to) where when trying to detach the database it was put into single-user mode. I'm wondering if that setting is cached somewhere and is the reason that SQL Server put it back into single-user mode after a failover.

    Read the article

  • Mirror your display to another computer on your network

    - by klikkamongo
    I have a computer I am using for showing media, running windows 7 (XBMC). It is hooked up with a projector. I want other computers on my network to have the opportunity to mirror their display to this computer.(without the media computer having to accept it, it should be all controlled by the sending computer.(like sending a movie with Apple's Airplay)) One scenario: I am finding a webpage with a great article (or a Word document) on my computer that i want to show to the other people in the room, i press a button, then my screen is mirrored to the media pc. I can still scroll, write etc. so everybody can see it. When i am done showing the article, I can exit the display mirroring, and the media computer display goes back to normal. This should mirror the whole screen, not only the browser like plugins can do in chrome, or sending Youtube movies like plugins for Opera. Update I found a video example of the solution I am looking for, unfortunately it looks like it will be hard making a Chromecast reciever for windows. Using the "Cast entire screen (experimentally)" button on the Chromecast plugin in Chrome is doing the exact thing I want to do, see this video.

    Read the article

  • Migration Guide: Migrating to SQL Server 2012 Failover Clustering and Availability Groups from Prior Clustering and Mirroring Deployments

    This paper provides guidance for customers who prior to SQL Server 2012 have deployed SQL Failover Clustering for local high availability and database mirroring for disaster recovery, and who want to migrate to SQL Server AlwaysOn. It describes the corresponding SQL Server AlwaysOn scenario and the migration paths to SQL Server AlwaysOn. It also contains the important knowledge and considerations that you must know in order to successfully migrate to a HADR solution based on SQL Server AlwaysOn technology, which implements AlwaysOn Failover Cluster Instances for high availability and AlwaysOn Availability Groups for disaster recovery.

    Read the article

  • Virtual Disk Degraded

    - by TheD
    There is a physical DC with a Raid 1 Mirror, 2 Physical Disks, 500GB each. Dell Server Administrator is installed on the DC, and is reporting both physical disks are fine, online, in a good state etc. On a PERC S300 Raid Controller: Physical Disk 0:0 Physical Disk 0:1 However at the same time it's reporting that a virtual disk is degraded, what exactly does this mean? The virtual disk indicates it's State is in a Raid 1 Layout. Device Name: Windows Disk 0 If my understanding is correct then the Virtual Disk, when you drill down into Dell OpenManage should have both physical disks as members, as it is a mirror? Is this correct? However, when I drill down into the Virtual Disk, it only displays Physical Disk 0:0 included in Virtual Disk 1. I'm very new to server side/raid management etc. just while our server techy is away! Thanks!

    Read the article

  • How to list my harddrives in OpenSolaris?

    - by Sanoj
    I have a machine with two harddrives. I have installed OpenSolaris on one of them and now I want to add the other one as a mirror-drive in my zpool. But to do that I have to know the name of my drive, something like c7d0s0. Using the command zpool status can I see the harddrives already in use in my zpool but not the hard-drives that are not in use. How can I list the names of the hard-drives that are not in use in any zpool or list the name of all my hard-drives?

    Read the article

  • Distributed Database Services?

    - by Cameron
    I'm working on a database-driven web service with clients in the US and Australia. We're currently hosted in the US, however our Australian clients are experiencing lag. The lag is primarily due to the fact that the pages launch AJAX queries which require some db work to be done on our database in the US and these take a while to perform a round trip. Ideally, we're looking for some kind of distributed database system which replicates our main US database in Australia (and possibly other locations if we choose to expand later on). Does anyone have any suggestions for services which offer something like this? Something like a CDN (CacheFly etc), which is web-based, simple to set up etc but for databases instead of static files. Ideally it would be completely transparent to the application and abstract away all the distributed database management, syncs etc.

    Read the article

  • secure synchronization of large amount of data

    - by goncalopp
    I need to automatically mirror a large amount (terabytes) of files in two unix machines over a slow link (1 Mbps). This needs to be done frequently, but the data doesn't change too much (delta transmission doesn't saturate the link). The usual solution would be rsync, but there's an additional requirement: it's undesirable, from a security standpoint, that either the source or destination machines have (keyless) ssh keys to each other, or any kind of filesystem access. All communication between the two machines should thus be initialized (and mediated) through a third machine. I've asked a separate question about rsync in particular here. Are there other obvious solutions I'm missing?

    Read the article

  • SQL 2008 Datawarehouse Collection Agent Fails with Mirrored Databases on instance

    - by Dale Wright
    I have a data collection job that fails when a Database on the instance is in Recovery Mode. The database in recovery is the MIRROR partner in the database mirror. The Job that fails is as follows. collection_set_1_noncached_collect_and_upload The job consists of the following steps dcexec -u -s 1 -i "INSTANCE03" EXEC [dbo].[sp syscollector purge collection logs] dcexec -u -s 1 -i "INSTANCE03" The job fails at Step 1. I have run the steps manually and they all appear to be ok. If I change the mirror database to be the principal the job completes successfully.

    Read the article

  • back-end SQL server 2005 databases for website

    - by Datapimp23
    Hi, We're migrating an existing IIS website + MS SQL 2005 database (on the same server) to a new test set-up. The existing set-up is too slow. I want one ISS server and 2 X MS SQL server 2005. One live DB server for the website queries (inserts, updates) and another for backups, reports or stored procedures. So the live DB should be more aimed at performance. The other doesn't even need to be synced instantly. What is the best way in SQL server 2005 to set this up. Can somebody point me in the right direction and give me some pointers. Thanks

    Read the article

  • NLB Accessed Denied from second server

    - by Igor K
    Hello Have NLB set up and working. Just two servers, web1 and web2. I set up NLB on web1 and can view both machines as Converged. (I had to specify a login/pass to connect to web2). On web2, looking at the NLB Manager, it only shows web2 and says in the log "Accessed Denied. Error connecting to web1". Any ideas how to fix?

    Read the article

  • How to add a mirror-disk to my OpenSolaris?

    - by Sanoj
    I have a machine with two harddrives. I have installed OpenSolaris on one of them and now I want to add the other one as a mirror-drive in my zpool rpool. I guess I have to format the second disk first and then add it to the pool. How can I do this?

    Read the article

  • How to Mirror or Clone a Spanned Volume in Windows 2008

    - by Matt
    I have a spanned volume (3x6+ TB disks spanned to one 20+ TB volume) that I need to mirror or clone to a new 20+ TB (unspanned) volume. Once mirrored or cloned I'm going to destroy the original volume and reuse the storage elsewhere. Windows 2008 will not allow me to mirror it because the original is a spanned volume. I cannot simply copy the data, because there are sparse files on the volume. So the OS thinks there is 150+ TB used on the disk when there really is only around 18TB used physically. When I try to use the copy command it won't run because it thinks the destination volume needs to be 150+ TB to hold it all. A conundrum, but I figure someone here has the answer. Thanks, Matt

    Read the article

  • Mesh Networked servers via vpn

    - by microspino
    I got a design idea and I would like to have some advice from SF about It. I have 5 customers with small real-estate databases. I've built for them a desktop app and now they would like to merge their database to share their data. I don't want to centralize everything in one place nor I want to do maintenance for servers. They told me also, that all of them in their offices, have little servers and maintenance guys available. Although everything seems suitable for web application, I had the idea to experiment something new: Any customer small-server wild be connected to the others in a sort of mesh network without a single point of failure and through VPNs. If one of the servers went down the customers could still connect to their databases from one of the other mesh networked servers instead of from the local one that is down. During normal operations all the servers sync the db with the others through VPNs. I can accept a half-day timing window of NON synched data, in other words, since I don't need real time synchronization, the server don't have to always stay in synch. I can migrate my data over to other Non-Sql technologies like CouchDB or Redis or whatever you suggest. As you can see I don't have a lot of constraints and although I could go with a web application I would like to delegate and decentralize support, data-privacy and management, as more as I can to my customers offices. Is that a crazy idea? Do you know If something similar exist? Which technology would you suggest?

    Read the article

  • Simplest way to shrink transaction log files on a mirrored production database

    - by MGOwen
    What's the simplest way to shrink transaction log file on a mirrored production database? I have to, as my disk space is running out. I will make a full database backup before I do this, so I don't need to keep anything from the transaction log (right? I have daily full database backup, probably never need point-in-time restore, though I'll keep the option open if I can - that's all the .ldf is really for, correct?). (Hope this isn't an exact duplicate, I read a lot of questions but couldn't find this exact scenario elsewhere).

    Read the article

  • Serve mirrored (static) web-page with original headers

    - by aioobe
    I have a dynamic webpage which I want to create a "frozen" copy of. Typically I would do something like wget -m http://example.com, and then put the files in the document root of the web-server. This site however has some dynamic content, including dynamically generated images, for instance http://example.com/company/123/logo This means that in order to mirror the page, I need to Save whatever headers the server currently serves for each URL. This can be done using the wget option --save-headers. Serve the static pages and serve the proper headers for each file. (This I have no idea of how to do.) What is the best way to solve this? Any suggestions are welcome.

    Read the article

  • wget recursively with -np option still ascends to parent directory

    - by vectra
    tl;dr: will `wget --no-parrent -r ' download from a directory above the given url's directory? when using wget to download, say images, recursively from example.com/a/b with the -r and -np options, will a picture that is under example.com/a/c/ be downloaded when example.com/a/b/ delivers a html-file containing a link to the picture? if so, how do i get all pictures, that are in a folder and it's subfolders and only those? the description of the option --no-parent says "Do not ever ascend to the parent directory when retrieving recursively". anyway directory browsing delivers a link to the parent directory, which wget will follow, despite mentioned option. now what did i miss? edit: using GNU Wget 1.12

    Read the article

  • Service or software to auto-sync one server to another so they are identical?

    - by Ryan
    We have a main node in a DC data center and want to setup a back-up node in Seattle. The back-up node will only be used if the DC node goes down and we switch it over while DC node is repaired. The question is, what kind of services out there allow me to sync the data, I suppose we want to do it fairly frequently so if something goes down there isn't much loss in data between the time of failure and the last back-up/sync. Is there any common solution for this? It's Windows Server 2003 running Parallels Virtuozzo.

    Read the article

  • How do I mirror a MySQL database?

    - by user45745
    I'm running two load balanced servers for one website, and I'd like the databases to be synchronized. Queries may be run on either of the two servers because they are both production sites, so the replication can't just work one way. It doesn't have to be in real-time, just fairly accurate so people don't notice a difference when they get switched to a different server.

    Read the article

  • Issue with broken disk on Solaris with raidctl - how to proceed

    - by weismat
    I have a SunFire T2000 server which has 2 mirrored disks pairs. The server required an exchange of the system battery. After swaping the battery first no disks were found. After booting from CD we managed to find the disks, but now one disk is broken and the raidctl reports a failed synchronisation. The boot process stops now when trying to mount the file systems. The power light of the broken drive is not even blinking. What is the best way to proceed now ? Fortunately I could live with loosing the data on the drive as it is backed up, but I would like to keep the rest of the data as it contains /etc and get the server booting again.

    Read the article

  • Why does my simple Raid 1 backup storage perform really slow sometimes?

    - by randomguy
    I bought 2x Samsung F3 EcoGreen 2TB hard disks to make a backup storage. I put them in Raid 1 (mirror) mode. Made a single partition and formatted it to NTFS, running Windows 7. For some reason, accessing the drive's contents (simply by navigating folders) is sometimes really slow. Like opening D:/photos/ can sometimes take several seconds before it starts showing any of the folder's contents. Same applies for other folders. What could be causing this and what could I do to improve the performance? I remember that there was an option somewhere inside Windows to choose fast access but less reliable persistence operations (read/write). It was a tick inside some dialog. At the time, it felt like a good idea to take the tick away from the option and get more reliable persistence but slower access, but now I'm regretting. I'm unable to find this dialog.. I've looked hard. I don't know, if it would make any difference. Oh, and I've ran scan disk and defrag on the drive. No errors and speed isn't improved.

    Read the article

  • Sync file system changes in a folder to server on LAN (via SSH or NFS) automatically

    - by Peter Farsinsen
    Hi, I'm looking for a solution that'll automatically sync file system changes in "/folder1" to "/folder2". Any change should be synced (deleted files, new folders etc.) without any confirmation. Preferably the changes will happen to "/folder2" instantly. "/folder1" is on my hdd "/folder2" is on a server on the LAN (can be accessed via SSH or NFS) I'm on OS X 10.6. I can't seem to find just the right tool. Does anyone

    Read the article

  • Download web server structure with empty files

    - by golimar
    I want to make a mirror of a Web server, but downloading the actual files will take too long. So I thought of having just the directory and file structure, and when I need the actual contents of the file, I can download just that file. I have tried wget --spider URL and in a short time it has created in my local disk the directory structure with no files. But I've checked all of wget's or curl's switches and there is nothing like what I need. Can this be done with wget, curl or any other tool?

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >