Search Results

Search found 62141 results on 2486 pages for 'alternate data stream'.

Page 807/2486 | < Previous Page | 803 804 805 806 807 808 809 810 811 812 813 814  | Next Page >

  • glusterfs to replicate files to other servers

    - by sbrattla
    I've got multiple servers which all need to have the same content in /home. In other words, if the file /home/user1/test.txt is updated on server A, this needs to be replicated to all other servers in the cluster. Is it possible to use GlusterFS for this purpose? That is, let each server have a full copy of all data locally - which that server will be working on - and solely use GlusterFS to take care of replicating this data to the other servers? I'm not intersted in a combined storage, but rather have all data on all machines only to have GlusterFS to replicate it to the other machines.

    Read the article

  • Is there an encrypted write-only file system for Linux?

    - by Grumbel
    I am searching for an encrypted filesystem for Linux that can be mounted in a write-only mode, by that I mean you should be able to mount it without supplying a password, yet still be able to write/append files, but neither should you be able to read the files you have written nor read the files already on the filesystem. Access to the files should only be given when the filesystem is mounted via the password. The purpose of this is to write log files or similar data that is only written, but never modified, without having the files themselves be exposed. File permissions don't help here as I want the data to be inaccessible even when the system is fully compromised. Does such a thing exist on Linux? Or if not, what would be the best alternative to create encrypted log files? My current workaround consists of simply piping the data through gpg --encrypt, which works, but is very cumbersome, as you can't easily get access to the filesystem as a whole, you have to pipe each file through gpg --decrypt manually.

    Read the article

  • How do I set up my own proxy server?

    - by NJTechGuy
    This website (abc.com) slowed access from our original IP address. How do I implement my own proxy server to hide my IP while browsing abc.com? Do I need special hardware/software combo to achieve this? If I can generate about 5 proxies and alternate amongst those 5 while browsing abc.com would be awesome. Please suggest. Thanks guys! p.s : I want to know if I can generate proxy IPs of the type 123.34.21.140 prot 80 on my own? I want to use those IP/port combos in my Python scripts (urllib2/set_proxy).

    Read the article

  • What does this WinImage Error Mean?

    - by Kryten
    Hi, I am using WinImage (version 8.0) on Windows XP SP3 to create a VHD File from my Hard Drive, however, everytime I try to, everytime it says: "I/O error in reading data on position 000000007cd40000:00010000. Error 23: Data Error (cyclic redundancy check)." with the options "Abort", "Retry", "Skip" & "Skip All". Everytime I try "Retry" the message comes, if I click skip another message comes up (same as above, except the data position). I have run "CHKDSK /F /R" several times, but the error always appears. What is wrong? Can I safely skip the Errors? Is there anything I can do to fix the error?

    Read the article

  • Can someone correct me about Windows 2008 DFS implementation

    - by cwheeler33
    I have 2 two Windows 2003 file servers using DoubleTake at the moment. They are in an 2003 AD domain. And each server has it's own disk set. It is time to replace the servers... I want to use Windows 2008 64bit Ent. I was thinking of using DFS-R to replace DoubleTake. The part I'm not sure about is clustering. Do I need to have shared storage if each server has a copy of the data? I want to have the data available to the same share name, so maybe I don't fully understand how DFS is set up. I currently have about 6TB of data. I expect to grow by 3TB a year on these file servers. Any resources/books that could teach me would be good to know as well.

    Read the article

  • How to forbid programs from accessing a folder

    - by prashanth
    I use firefox as my default browser. IE and chrome are used as secondary browsers. Apart from these, i need to have a alternate browser. I tried opera, but when installing it, opera automatically imported all my bookmarks and other profile data from firefox folder. I really had a hard time safely deleting these files that opera made a copy, without harming the original files which firefox uses. I don't want this to happen. Here's the situation: Now, i want to install other browser or re install opera. How do i make the firefox profile folder to be accessed only by firefox.exe and me(the administrator)? Can this be done via security tab under folder properties?

    Read the article

  • LAN access via USB from Ipod Touch?

    - by Alec
    I need to browse the local web server from my Ipod Touch to test apps we're developing. I'm not allowed to install a separate wireless access point which would be the easiest solution. Can I use the USB cable for this? Also, the local PC is a Dell Mini 9 running Ubuntu. Has anyone managed to use the wireless port to create an ad hoc connection to an Ipod Touch so the Ipod can browse the Ubuntu web server? This would be an alternate option for me. Thank you!

    Read the article

  • Make Exchange 2007 use the correct SSL certificate

    - by Neil
    I have an SBS 2008 server contososerver.contosodomain.local which is externally accessible with the domain remote.contoso.com and an SSL certificate for the external domain which we installed using the SBS 2008 wizard. This works great for OWA because IIS serves the remote.contoso.com certificate. I also want to turn on external POP3/IMAP4/SMTP however when I try, I get served the internal certificate that SBS generated automatically (using its internal CA) which has the alternate names remote.contoso.com, contososerver.contosodomain.local and contososerver. I tried removing this certificate from Exchange but it won't let me because it needs it for its internal receive connector. So how do I tell Exchange 2007 to use the real certificate for external POP3/IMAP4/SMTP?

    Read the article

  • Convert Public Folders to a PST

    - by TrueDuality
    Alrighty so I've got a tricky one. I currently have a public folder database (edb & stm) residing on an Exchange 2003 folder. I need to export them into a pst file or otherwise make it so that I can manually get the data in it to end-users. I can not use the export feature built into Outlook as some of the folder refer to another server which doen't have the data. Trying only results in the Outlook Client hanging for close to an hour before giving an error about not finding the data. So this will need to be a server side export. There are a few tools out there that seem to be available for converting edb & stm files to psts but they are quite expensive. Does anybody have any ideas?

    Read the article

  • Securing a persistent reverse SSH connection for management

    - by bVector
    I am deploying demo Ubuntu 10.04 LTS servers in environments I do not control and would like to have an easy and secure way to administer these machines without having to have the destination firewall forward port 22 for SSH access. I've found a few guides to do this with reverse port (e.g. howtoforge reverse ssh tunneling guide) but I'm concerned with security of the stored ssh credentials required for the tunnel to be opened automatically. If the machine is compromised (primary concern is physical access to the machine is out of my control) how can I stop someone from using the stored credentials to poke around in the reverse ssh tunnel target machine? Is it possible to secure this setup, or would you suggest an alternate method?

    Read the article

  • What are best monitoring tool customizable for cluster / distributed system?

    - by Adil
    I am working on a system having multiple servers. I am interested in monitoring some server specific data like CPU/memory usage, disk/filesystem usage, network traffic, system load etc. and some other my process specific data. What are available open source that can serve my purpose? If it provides to customize the parameter to be monitored and monitor your own data by creating plugin / agent. Any suggestions? I heard of Nagios, Zabbix and Pandora but not sure if they provide such interface.

    Read the article

  • Is Family Tree Maker 2011 the right upgrade?

    - by bill weaver
    My father has used Family Tree Maker for years, but hasn't upgraded since version 11. It is difficult to tell from reviews at Amazon and other places whether upgrading to FTM 2011 is a good choice. File incompatibilities and upgrade woes sound like customer service is lacking, and i've read reports of it uploading your data to their database but then trying to sell you a download of data. Looking at the ancestry.com site makes me think it's solely about selling add-ons and upgrades. On the other hand, the feature set seems fairly rich and the software has a pretty strong following. I was able to get Gramps working on my system, but that's not going to work for my dad. Any advice on a good upgrade path? Doesn't necessarily have to be FTM. The only requirement is a way to import his existing data.

    Read the article

  • Sudden loss of Wi-Fi connectivity on OS X

    - by GJ.
    Occasionally while I work, without any special provocation, I lose connectivity via Wi-Fi. Other devices connected to the same wifi network have no interruption, and the problem gets resolved once I reboot my MacBook Air, so it's definitely a local problem. Observations: The Wi-Fi symbol in the menu indicates that I'm still connected, but apps can't actually connect neither to the Internet nor to other devices in the LAN. I can't connect to an alternate Wi-Fi network (e.g. Wi-Fi tethering via iPhone). I can connect to the Internet via iPhone USB tethering but this seems to only work some of the time. Only a reboot solves the problem but a regular restart gets stuck on a grey screen with rotating wheel (after all applications have closed) and I have to do a hard reset. How should I go about troubleshooting this? It used to happen very rarely but now is becoming more frequent (approaching once every 2-3 days on average).

    Read the article

  • Snow Leopard, PHP, and MySQL

    - by Peter
    I have just installed Snow Leopard and now my PHP/MySQL program ends in a "Segmentation fault". I have been searching the web for a solution, I realize that there are some issues with SL/PHP/MySQL, but I have not found anything that works yet. I downloaded the binary MySQL package mysql-5.1.42-osx10.6-x86_64. I have updated the php.ini file as suggested at various posts. When I run PHP and connect to the MySQL server the behavior is a bit random. In many cases it works fine to connect and read data. In my specific case the PHP program constructs a LOAD DATA LOCAL INFILE ... statement to load data from a text file. It should do several such queries after each other in a loop. It works one time but the halts in a "Segmentation fault". The program worked fine in Leopard, but not now. My versions are: OS 10.6.2, PHP 5.3.0, MySQL 5.1.42

    Read the article

  • lamp server permissions on development server

    - by user101289
    I run a LAMP server on a ubuntu laptop I use only for development. I am not greatly concerned with security, since the server is never accessible outside the local network, and it's turned off when I'm not using it. My question is what is the simplest and 'best' way to set permissions/users/groups so that when my myself user creates, edits or writes files in the webroot, I won't need to go through and CHMOD / CHOWN everything back to the www-data user? Should I add myself to the www-data group? Or chown the webroot to www-data:myself? Or is there a best practice for this situation so I don't have to keep re-setting the ownership of these files? Thanks

    Read the article

  • Concurrent modification during backup: rsync vs dump vs tar vs ?

    - by pehrs
    I have a Linux log server where multiple applications write data. Data is written in bursts, and in a lot of different files. I need to make a backup of this mess, preferably preserving as much coherence between the file versions as possible and avoiding getting truncated files. Total amount of data on the server is about 100Gb. What I really would want (but can't) is to shut-down, backup the system cold and then start it up again. What kind of guarantees against concurrent modification does the various backup tools give? When do they "freeze" the file versions? I am looking at rsync, dump and tar at the moment, but I am open for other (open source) alternatives. Changing the application or blocking writing for backups is sadly not an option. System is not running LVM (yet), but I have considered that for rebuilding the system and then snapshots.

    Read the article

  • how to restore grub

    - by takeshin
    During copying files via lan my Ubuntu Server 10.04 hung up and I had to reset the computer. After reboot, I got: Grub error 17, so I tried the rescue alternate CD, but I got a info, that no partitions were found on the disk. I used testdisk to restore the partitions. Using fdisk -l shows the partitions now, but when the system boots up I does nothing but displays: L234: When I plug the drive to another computer, it is not automatically mounted and I can't access the data. What to do now? How to restore the grub and boot up the system?

    Read the article

  • Speed up file access on home network

    - by kurasa
    I have 2 PCs (Windows 7 Ultimate) and a Mac running Windows 7 using vmware fusion on my home network tied together using WRN1000 NETGEAR Router On one of the PC's I have a set of file (MYOB .myo). These use a data source to access the data in the files. Operations (reading,writing) to the .myo on the PC which hosts the files is fine but the other 2 it is painfully slow/unreliable and I am wondering what I can do to speed this up. Some ideas I have are 1. Turn off the Windows firewall on all the windows installations on the home network 2. Buy another router. Specifically a router which I can connect a USB flash drive on the back where I can put the .myo files and all the PC can access the files from the USB flash drive on the router (does this speed things up?) Any advice greatly appreciated on how I can speed up this access to data

    Read the article

  • All Tasks..Advanced Operations option missing from Certificaces MMC Snap-In

    - by JohnFx
    I am trying to follow the instructions in this article to create a custom certificate to support SSL on a web server. I'm stick on the following step: Click on Personal – All Tasks – Advanced Operations – Create Custom request The problem is that on the web server (Windows Server 2003 R2) I don't have an "Advanced Operations" option under "All Tasks". I do on my desktop machine (Windows 7), but not on the server. All the documentation I can find indicates that it should be available on WS-2003-R2, but it just isn't. Note: I'm going through this manual process because I need to specify a alternate host names in the CSR, which you can't do through the IIS 6.0 console certificate managment functionality. Any suggestions for how to make this option show up?

    Read the article

  • How to efficiently dump a huge MySQL innodb database?

    - by Jagbir
    I got an Ubuntu 10.04 production MySQL database server where total size of database is 260 GB while size of root partition is itself 300 GB where DB is stored, essentially means around 96% of / is full and there's no space left for storing dump/backup etc. No other disk is attached to server as of now. My task is to migrate this database to other server sitting in different datacenter. Question is how to do that efficiently with minimum downtime? I'm thinking in line of: Request to attach an extra drive to server and take a dump in that drive. Transfer dump to new server, restore it and make new server slave of existing one to keep data in sync When migration is needed, break replication, update slave config to accept read/write requests and make old server read-only so it won't entertain any write requests and tell app developers to update there config with new IP address for db. What's your suggestions to improve this or any alternate better approach for this task?

    Read the article

  • USB Flash drive corrupt files - Help needed

    - by Michael
    I have a 16gb usb flash drive with 8gb worth of data that I can't afford to lose. When I inserted it into my pc, inside the folder that I was storing the data I saw unintelligible characters and nothing in there would open. I ran, windows scandisk and the files (unfortunately) disappeared. I can see that the drive's space still appears to be taken up with data, about 8gb. What should I do to recover it? Is it possible? Thanks in advance... Michael

    Read the article

  • Upgrading from 32 to 64 Bit Windows 7, without losing program installations/games

    - by Fogest
    I recently built a new computer and put the wrong installation of Windows on (32 bit), meaning I cannot use all of my RAM. I would like to upgrade to the 64 bit version, though I already have downloaded many programs and games which would total to around 30 GB give or take. I don't have the kind of data usage with my ISP to re-download this much data again, until next month (total GB will be higher as time goes on). I know there is Windows Easy Transfer, but it is not so much my data itself I'm worried about, it is more having to re-download and install a bunch of games and applications. Is it possible to perform an upgrade from 32 bit to 64 bit without this loss?

    Read the article

  • IIS needs to be restarted every morning

    - by Kevin
    In one of my Application and DB Server , SSIS package runs at night. Every morning i need to reset IIS to work the Application Fast and smoothly. One day i tried to SKIP the SSIS Package and next day i hvnt Done the IIS reset. What could be the problem. Is there any alternate Solution for IIS reset. How can i schedule and make sure the IIS is RESTARTED through Batch File / s. Application is developed in .NET and DB is SQL latest version. The application is hostes on cloud server. Your prompt reply will be helpful for me.

    Read the article

  • Is it possible to FORMAT an external hard disk that has been encrypted using Storagecrypt?

    - by Pandian John
    Basically the big problem is that about 680 GB of data from my Seagate 2 TB Ext HD is lost because I was experimenting with a software called storagecrypt. I used it a few months ago and today I tried it again but i didn't know that the old password is already set in the hard disk when I pressed the encrypt button. I forgot the password which is disappointing. Not to mention that software uses 128 bit AES encryption so there is no way iam going to recover that data. My question is that is it possible to Format my Hard disk which has been encrypted? What i mean is that is it possible to completely wipe the data just like it is newly bought so that I can use my External Hard disk?( I tried to format by right click-- Format. But the size of the disk is shown as 1 MB. Answers would be very much appreciated. Thanks.

    Read the article

  • breaking mdadm raid and moving to NTFS

    - by daveyt
    I'm running Ubuntu 8 something and my data is on a mirrored pair of 1TB disks formatted as ext3, and the RAID is via mdadm. I want to move to Windows 7 (yeah yeah I know but Linux aint doing it for me at the moment) and migrate the disks to NTFS. My plan is: Break the MDADM RAID (by failing one disk logically) Format the 'failed' disk as NTFS Copy data from the RAID array to the NTFS disk (dont care about perms) Install Windows, (new separate non RAid disk) and my data disk is available. I've researched this and it seems the easiest way. I dont have another disk to back up to so I think this is my only way. Can anyone see a better/easier way?

    Read the article

< Previous Page | 803 804 805 806 807 808 809 810 811 812 813 814  | Next Page >