Search Results

Search found 34699 results on 1388 pages for 'database backup'.

Page 28/1388 | < Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >

  • rdiff-backup is taking longer and longer every time it runs

    - by Jakobud
    I've been running rdiff-backup for the past week or so, every night at 4am. It started out not taking that long, but its taking longer and longer and longer every time its runs every night. In some cases there are quite a few new and changed files and in other cases, not so much. It started out taking < 10 minutes, and after a week its taking over 4.5 hrs to run. Take a look at my session stats below: StartTime 1268046002.00 (Mon Mar 8 04:00:02 2010) EndTime 1268046373.50 (Mon Mar 8 04:06:13 2010) ElapsedTime 371.50 (6 minutes 11.50 seconds) SourceFiles 213928 SourceFileSize 277271282225 (258 GB) MirrorFiles 213914 MirrorFileSize 276693097638 (258 GB) NewFiles 16 NewFileSize 578209911 (551 MB) DeletedFiles 2 DeletedFileSize 2598 (2.54 KB) ChangedFiles 27 ChangedSourceSize 5195150 (4.95 MB) ChangedMirrorSize 5217876 (4.98 MB) IncrementFiles 0 IncrementFileSize 0 (0 bytes) TotalDestinationSizeChange 578184587 (551 MB) Errors 0 StartTime 1268132402.00 (Tue Mar 9 04:00:02 2010) EndTime 1268134341.29 (Tue Mar 9 04:32:21 2010) ElapsedTime 1939.29 (32 minutes 19.29 seconds) SourceFiles 213963 SourceFileSize 307959842562 (287 GB) MirrorFiles 213928 MirrorFileSize 277271282225 (258 GB) NewFiles 37 NewFileSize 31265005547 (29.1 GB) DeletedFiles 2 DeletedFileSize 576511960 (550 MB) ChangedFiles 25 ChangedSourceSize 5243761 (5.00 MB) ChangedMirrorSize 5177011 (4.94 MB) IncrementFiles 65 IncrementFileSize 577266412 (551 MB) TotalDestinationSizeChange 31265826749 (29.1 GB) Errors 0 StartTime 1268218802.00 (Wed Mar 10 04:00:02 2010) EndTime 1268225230.15 (Wed Mar 10 05:47:10 2010) ElapsedTime 6428.15 (1 hour 47 minutes 8.15 seconds) SourceFiles 213971 SourceFileSize 307960643843 (287 GB) MirrorFiles 213963 MirrorFileSize 307959842562 (287 GB) NewFiles 9 NewFileSize 694087 (678 KB) DeletedFiles 1 DeletedFileSize 894 (894 bytes) ChangedFiles 38 ChangedSourceSize 30656167797 (28.6 GB) ChangedMirrorSize 30656059709 (28.6 GB) IncrementFiles 48 IncrementFileSize 289278151 (276 MB) TotalDestinationSizeChange 290079432 (277 MB) Errors 0 StartTime 1268305202.00 (Thu Mar 11 04:00:02 2010) EndTime 1268312788.15 (Thu Mar 11 06:06:28 2010) ElapsedTime 7586.15 (2 hours 6 minutes 26.15 seconds) SourceFiles 213971 SourceFileSize 307960643779 (287 GB) MirrorFiles 213971 MirrorFileSize 307960643843 (287 GB) NewFiles 0 NewFileSize 0 (0 bytes) DeletedFiles 0 DeletedFileSize 0 (0 bytes) ChangedFiles 15 ChangedSourceSize 30650824127 (28.5 GB) ChangedMirrorSize 30650824191 (28.5 GB) IncrementFiles 16 IncrementFileSize 689437042 (657 MB) TotalDestinationSizeChange 689436978 (657 MB) Errors 0 StartTime 1268391601.00 (Fri Mar 12 04:00:01 2010) EndTime 1268400145.85 (Fri Mar 12 06:22:25 2010) ElapsedTime 8544.85 (2 hours 22 minutes 24.85 seconds) SourceFiles 213974 SourceFileSize 307960784445 (287 GB) MirrorFiles 213971 MirrorFileSize 307960643779 (287 GB) NewFiles 5 NewFileSize 269587 (263 KB) DeletedFiles 2 DeletedFileSize 128921 (126 KB) ChangedFiles 16 ChangedSourceSize 30650815948 (28.5 GB) ChangedMirrorSize 30650815948 (28.5 GB) IncrementFiles 26 IncrementFileSize 818279963 (780 MB) TotalDestinationSizeChange 818420629 (781 MB) Errors 0 StartTime 1268478002.00 (Sat Mar 13 04:00:02 2010) EndTime 1268488740.33 (Sat Mar 13 06:59:00 2010) ElapsedTime 10738.33 (2 hours 58 minutes 58.33 seconds) SourceFiles 213974 SourceFileSize 307960784238 (287 GB) MirrorFiles 213974 MirrorFileSize 307960784445 (287 GB) NewFiles 0 NewFileSize 0 (0 bytes) DeletedFiles 0 DeletedFileSize 0 (0 bytes) ChangedFiles 16 ChangedSourceSize 30650823920 (28.5 GB) ChangedMirrorSize 30650824127 (28.5 GB) IncrementFiles 19 IncrementFileSize 1041846015 (994 MB) TotalDestinationSizeChange 1041845808 (994 MB) Errors 0 StartTime 1268560803.00 (Sun Mar 14 04:00:03 2010) EndTime 1268573194.82 (Sun Mar 14 07:26:34 2010) ElapsedTime 12391.82 (3 hours 26 minutes 31.82 seconds) SourceFiles 213974 SourceFileSize 307960784238 (287 GB) MirrorFiles 213974 MirrorFileSize 307960784238 (287 GB) NewFiles 0 NewFileSize 0 (0 bytes) DeletedFiles 0 DeletedFileSize 0 (0 bytes) ChangedFiles 14 ChangedSourceSize 30650815948 (28.5 GB) ChangedMirrorSize 30650815948 (28.5 GB) IncrementFiles 15 IncrementFileSize 1116911810 (1.04 GB) TotalDestinationSizeChange 1116911810 (1.04 GB) Errors 0 StartTime 1268647203.00 (Mon Mar 15 04:00:03 2010) EndTime 1268662364.88 (Mon Mar 15 08:12:44 2010) ElapsedTime 15161.88 (4 hours 12 minutes 41.88 seconds) SourceFiles 214010 SourceFileSize 307963430178 (287 GB) MirrorFiles 213974 MirrorFileSize 307960784238 (287 GB) NewFiles 37 NewFileSize 2684172 (2.56 MB) DeletedFiles 1 DeletedFileSize 5348 (5.22 KB) ChangedFiles 32 ChangedSourceSize 30656134913 (28.6 GB) ChangedMirrorSize 30656167797 (28.6 GB) IncrementFiles 71 IncrementFileSize 1316460362 (1.23 GB) TotalDestinationSizeChange 1319106302 (1.23 GB) Errors 0 StartTime 1268733603.00 (Tue Mar 16 04:00:03 2010) EndTime 1268750396.76 (Tue Mar 16 08:39:56 2010) ElapsedTime 16793.76 (4 hours 39 minutes 53.76 seconds) SourceFiles 214010 SourceFileSize 307963430156 (287 GB) MirrorFiles 214010 MirrorFileSize 307963430178 (287 GB) NewFiles 0 NewFileSize 0 (0 bytes) DeletedFiles 0 DeletedFileSize 0 (0 bytes) ChangedFiles 15 ChangedSourceSize 30650823898 (28.5 GB) ChangedMirrorSize 30650823920 (28.5 GB) IncrementFiles 16 IncrementFileSize 936032413 (893 MB) TotalDestinationSizeChange 936032391 (893 MB) Errors 0 StartTime 1268820002.00 (Wed Mar 17 04:00:02 2010) EndTime 1268834619.90 (Wed Mar 17 08:03:39 2010) ElapsedTime 14617.90 (4 hours 3 minutes 37.90 seconds) SourceFiles 214010 SourceFileSize 307963430156 (287 GB) MirrorFiles 214010 MirrorFileSize 307963430156 (287 GB) NewFiles 0 NewFileSize 0 (0 bytes) DeletedFiles 0 DeletedFileSize 0 (0 bytes) ChangedFiles 14 ChangedSourceSize 30650815948 (28.5 GB) ChangedMirrorSize 30650815948 (28.5 GB) IncrementFiles 15 IncrementFileSize 1289272860 (1.20 GB) TotalDestinationSizeChange 1289272860 (1.20 GB) Errors 0 Is this common behavior? It's only about 300gigs of data total. One important thing to note though, rdiff-backup is backup up stuff over to a USB2 external drive. Maybe thats why its so slow?

    Read the article

  • Announcing Oracle Audit Vault and Database Firewall

    - by Troy Kitch
    Today, Oracle announced the new Oracle Audit Vault and Database Firewall product, which unifies database activity monitoring and audit data analysis in one solution. This new product expands protection beyond Oracle and third party databases with support for auditing the operating system, directories and custom sources. Here are some of the key features of Oracle Audit Vault and Database Firewall: Single Administrator Console Default Reports Out-of-the-Box Compliance Reporting Report with Data from Multiple Source Types Audit Stored Procedure Calls - Not Visible on the Network Extensive Audit Details Blocking SQL Injection Attacks Powerful Alerting Filter Conditions To learn more about the new features in Oracle Audit Vault and Database Firewall, watch the on-demand webcast.

    Read the article

  • Oracle Database 11g Helps Control Exponential Data Growth

    - by [email protected]
    The 2010 ESG annual customer survey is now available. As part of it, ESG interviewed 300 customers about their IT priorities and, unsurprisingly, "Manage Data Growth" is top of the list. Perhaps less self-evident is the proposed solution to target this prime concern: "Often overlooked because it is a database platform, Oracle Database 11g offers additional capabilities such as automatic storage management (ASM), advanced data compression, and data protection that make managing data growth much easier for organizations of any size." The paper goes on to discuss these capabilities and highlights their potential benefits. Oracle Database 11g Helps Control Exponential Database Growth - a worthwhile read for anyone having to deal with rapidly increasing amounts of data. Download your free copy here.

    Read the article

  • How does I/O work for large graph databases?

    - by tjb1982
    I should preface this by saying that I'm mostly a front end web developer, trained as a musician, but over the past few years I've been getting more and more into computer science. So one idea I have as a fun toy project to learn about data structures and C programming was to design and implement my own very simple database that would manage an adjacency list of posts. I don't want SQL (maybe I'll do my own query language? I'm just having fun). It should support ACID. It should be capable of storing 1TB let's say. So with that, I was trying to think of how a database even stores data, without regard to data structures necessarily. I'm working on linux, and I've read that in that world "everything is a file," including hardware (like /dev/*), so I think that that obviously has to apply to a database, too, and it clearly does--whether it's MySQL or PostgreSQL or Neo4j, the database itself is a collection of files you can see in the filesystem. That said, there would come a point in scale where loading the entire database into primary memory just wouldn't work, so it doesn't make sense to design it with that mindset (I assume). However, reading from secondary memory would be much slower and regardless some portion of the database has to be in primary memory in order for you to be able to do anything with it. I read this post: Why use a database instead of just saving your data to disk? And I found it difficult to understand how other databases, like SQLite or Neo4j, read and write from secondary memory and are still very fast (faster, it would seem, than simply writing files to the filesystem as the above question suggests). It seems the key is indexing. But even indexes need to be stored in secondary memory. They are inherently smaller than the database itself, but indexes in a very large database might be prohibitively large, too. So my question is how is I/O generally done with large databases like the one I described above that would be at least 1TB storing a big adjacency list? If indexing is more or less the answer, how exactly does indexing work--what data structures should be involved?

    Read the article

  • Inserting data to database from Android

    - by Angel
    I have to build an application where the requirement is that my clients will send data from their Android device and I have to save that data to a database. I have done the part of coding that inserts data from Android emulator to my XAMPP database on localhost, now I have to implement the real thing. How can I connect the devices where my application will be installed to the XAMPP database I have created so that the data they send can be inserted into it?

    Read the article

  • Building a database class in PHP

    - by Sprottenwels
    I wonder if I should write a database class for my application, and if so, how to accomplish it? Over there on SO, a guy mentioned it should be written as an abstract class. However, I can't understand why this would be a benefit. Do I understand correctly, that if I would write an abstract class, every other class that methods will need a database connection, could simply extend this abstract class and have it's own database object? If so, how is this different from a "normal" class where I could instantiate an database object? Another method would be to completely forget about my own class and to instantiate a mysqli object on demand. What do you recommend?

    Read the article

  • Join the Webcast on September 18 to Learn Benefits of Upgrading to Oracle Database 11g

    - by Cinzia Mascanzoni
    Attend the Webcast on Tuesday September 18, 2012, at 10 a.m. PT to learn the "Three Compelling Reasons to Upgrade to Oracle Database 11g." During the live Webcast, Oracle experts will explain how customers who are still working with Oracle Database 10g or an even older version can gain the business, operational, and technical benefits provided by Oracle Database 11g. If you cannot participate in the live event, a replay will be available on the same registration page shortly afterward.

    Read the article

  • The Future of the Database Begins

    - by Thanos Terentes Printzios
    For more than three-and-a-half decades, Oracle has defined database innovation. With our leading technologies, Oracle customers have been able to out-think and out-perform their competition. Soon organizations will be able to do that even faster.With the introduction of the Oracle Database In-Memory Option it will be possible to perform TRUE real-time, ad-hoc, analytic queries on your organization’s business data as it exists at that moment and receive the results immediately. Imagine your sales team being able to know the total sales they have made as of right now -- not last week, or even last night, but right now.Imagine innovation that accelerates business decision making to real-time speeds. That's the power of Oracle Database In-Memory.Watch Larry Ellison to find out what this and the other new features of Oracle Database 12c will do for you. Register Now for the Live Webcast

    Read the article

  • save data in the database to xml in c [closed]

    - by Jayanth N
    I have some data in the database. I want those data in database to be stored as an xml file. I'm using postgresql 9.1 for database, for xml processing I'm using libxml (http://xmlsoft.org/). I'm writing the code in C language. Please help me. Detailed explanation: I have a client, which sends me a xml file. Server receives the xml file, parses the xml file and stores it in the db. From db i want to send the details in the form of an xml to the client. client: <employee> <name>glen</name> <telephone>123456789</telephone> </employee> <employee> <name>gwen</name> <telephone>123456789</telephone> </employee> server parses this xml file as displayed below: name : glen telephone:123456789 name : gwen telephone: 123456789 and saves it in a database(postgresql9.1) if the client requests for details of the employees, i've to send it in xml form from database.I don't know how to do it can u help me out.

    Read the article

  • Database Delivery Patterns and Practices

    Continuous database delivery is an automated process for building, deploying and testing databases to reduce risk and make rapid releases possible. It's enabled by a pipeline that starts when database changes are checked in, and ends when they're deployed to production. The articles collected here will help you understand the theories and methodologies behind every stage of the database delivery pipeline.

    Read the article

  • How to Map a CSV or Tab Delimited File to MySQL Multi-Table Database [migrated]

    - by Keefer
    I've got a pretty substantial XLS file a client provided 830 total tabs/sheets. I've designed a multi table database with PHPMyAdmin (MySQL obviously) to house the information that's in there, and have populated about 5 of those sheets by hand to ensure the data will fit into the designed database. Is there a piece of software or some sort of tool that will help me format this XLS document and map it to the right places in the database?

    Read the article

  • Cron job failing to backing up a Postgres database

    - by user705142
    I'm unsure what's going on here: I've got a backup script which runs fine under root. It produces a 300kb database dump in the proper directory. When running it as a cron job with exactly the same command however, an empty gzip file appears with nothing in it. The cron log shows no error, just that the command has been run. This is the script: #! /bin/bash DIR="/opt/backup" YMD=$(date "+%Y-%m-%d") su -c "pg_dump -U postgres mydatabasename | gzip -6 > "$DIR/database_backup.$YMD.gz" " postgres # delete backup files older than 60 days OLD=$(find $DIR -type d -mtime +60) if [ -n "$OLD" ] ; then echo deleting old backup files: $OLD echo $OLD | xargs rm -rfv fi And the cron job: 01 10 * * * root sh /opt/daily_backup_script.sh It produces a database_backup file, just an empty one. Anyone know what's going on here?

    Read the article

  • Oracle Database Appliance:???????????1Box?????2????????!

    - by Yusuke.Yamamoto
    11?14????????·????????Oracle Database Appliance???????????????? ????????:????????Oracle Database Appliance??????????? Oracle Database Appliance ??? Oracle Database Appliance ??Oracle Database ?????????????????????????·??????????Oracle Database ??(1)??????(2)RAC One Node ??(3)Oracle RAC ?????????????????????? Oracle Real Application Clusters(RAC)|??????????? ??????Oracle Database 11gR2 Oracle Real Application Clusters One Node ??(1)?????DB?????????????1Box???????? Oracle Database Appliance ???Oracle Real Application Clusters(RAC) ????????????DB??????????????????(????2??????????????????????????????????????????)?1Box????4U???????????????????????? ??(2)?????DB????2???????? Oracle Database Appliance ???Oracle Appliance Manager ????????????????????????Oracle Appliance Manager ????????(7????)???????????(Oracle Database?Oracle Grid Infrastructure?Oracle Enterprise Manager??)?????????(????????????????)?????????????????????2??????? ???Oracle Database Appliance ???????????????????????·????????????? Oracle Appliance Manager:????????????:7??????????????????? ??(3)????????????:???CPU???????????????????? Oracle Database Appliance ????Pay-As-You-Grow(?????????????)???????????????·????????????????Oracle Database Enterprise Edition ???????2??~24????????????? ?????????????????Oracle Database Enterprise Edition ????????(??????????????????)??????????????? Oracle Database Appliance:???? ????????????????????????????????? Oracle Database Appliance:???? ?????? Oracle Database Appliance Oracle Database Appliance:?????? Oracle Database Appliance:?????(??) Oracle Database Appliance:3D?? ????????? Oracle Direct ????Oracle Appliance Manager ????????????????????

    Read the article

  • How to Link VS2010 Database Project and LINQ to SQL

    - by Jason
    As I am working with the new database projects in VS2010, and as I am learning LINQ to SQL, I am curious as to the best way to link the two groups of information so that when I update one, the other updates along with it. From my research here at SO, as well as in Google, it appears the general rule of thumb is: "Build the database, and then create your LINQ to SQL classes." Of course, if I make a change in my database, the LINQ to SQL doesn't update automatically and I have to do it by hand. This is fairly simple right now as my database is small, but I am curious if there is an easier way for this to happen. In addition, the LINQ to SQL tool is pretty nice. The ability to create tables, add associations, and even create inheritance is very simple. As my second question, I am curious as to whether or not VS2010 can work the other way - I design the database in the DBLM file and then link it back to my database project. I appreciate any help with either of these two questions. I'm really interested in making this as easy as possible to reduce errors during development and improve the speed at which changes can be made.

    Read the article

  • How to backup database (MS SQl Server 2008) in C# without using SMO (having proplems) ?

    - by SzamDev
    Hi I have this code and it is not working but I don't why? try { saveFileDialog1.Filter = "SQL Server database backup files|*.bak"; saveFileDialog1.Title = "Database Backup"; if (saveFileDialog1.ShowDialog() == DialogResult.OK) { SqlCommand bu2 = new SqlCommand(); SqlConnection s = new SqlConnection("Data Source=M1-PC;Initial Catalog=master;Integrated Security=True;Pooling=False"); bu2.CommandText = String.Format("BACKUP DATABASE LA TO DISK='{0}'", saveFileDialog1.FileName); s.Open(); bu2.ExecuteNonQuery(); s.Close(); MessageBox.Show("ok"); } } catch (Exception ex) { MessageBox.Show(ex.ToString()); } and I get this error : What is the proplem?

    Read the article

  • Why use SQL database?

    - by martinthenext
    I'm not quite sure stackoverflow is a place for such a general question, but let's give it a try. Being exposed to the need of storing application data somewhere, I've always used MySQL or sqlite, just because it's always done like that. As it seems like the whole world is using these databases, most of all software products, frameworks, etc. It is rather hard for a beginning developer like me to ask a question - why? Ok, say we have some object-oriented logic in our application, and objects are related to each other somehow. We need to map this logic to the storage logic, so we need relations between database objects too. This leads us to using relational database and I'm ok with that - to put it simple, our database rows sometimes will need to have references to other tables' rows. But why do use SQL language for interaction with such a database? SQL query is a text message. I can understand this is cool for actually understanding what it does, but isn't it silly to use text table and column names for a part of application that no one ever seen after deploynment? If you had to write a data storage from scratch, you would have never used this kind of solution. Personally, I would have used some 'compiled db query' bytecode, that would be assembled once inside a client application and passed to the database. And it surely would name tables and colons by id numbers, not ascii-strings. In the case of changes in table structure those byte queries could be recompiled according to new db schema, stored in XML or something like that. What are the problems of my idea? Is there any reason for me not to write it myself and to use SQL database instead?

    Read the article

  • Synchronize a client database with the central database

    - by Pavan Kumar
    I need to update existing data or insert new data from client database say DB1 into central database say DB2 both holding same schema and both databases reside in same machine. The updates are not biderectional. I just want changes to be reflected from client(DB1) to server(DB2). How do i achieve this using C# .NET ? Can anyone provide an example ?

    Read the article

  • Database tables - how many database?

    - by Thomas
    How many databases are needed for a social website? I have my tech team working on developing a social site but all their tables are in 1 database. I wanted to create separate table sets for user data, temporary tables, etc and thinking maybe have one separate database only for critical data, etc but I am not a tech person and now sure how this works? The site is going to be a local reviews website.

    Read the article

  • Database design for heavy timed data logging

    - by user293995
    Hi, I have an application where I receive each data 40.000 rows. I have 5 million rows to handle (500 Mb MySQL 5.0 database). Actually, thoses rows are stored in the same table = slow to update, hard to backup, ... Which king of scheme is used in such application to allow long term accessibility to the datas without problems with too big tables, easy backup, fast read / write ? Is postgresql better than mysql for such purpose ? Thanks in advance BEst regards

    Read the article

  • Diagnosing permission problems with Cobian Backup to network share

    - by DaveBurns
    I'm running the latest Cobian 11. I have a Synology DS412 NAS. All of my machines (Mac and Windows) access this just fine when I'm logged in and I browse to it manually. I have Cobian installed as a service on two Windows machines: WinXP SP3 and Win7 x64. On both machines, the service is set to log on with my user account which is in the Windows administrator group. Backups on both machines fail with the message "Couldn't create the destination directory "\nas1\backups\foo\bar\": The filename, directory name, or volume label syntax is incorrect". I have tried setting the NAS's share to allow anonymous read/write access but it made no difference. Although I want the backups to run unattended in the middle of the night, I have tested them by running them manually while I'm logged in but no luck. Before starting that, I make sure that I can browse to the NAS with Explorer to ensure that any authentication session with Windows and the NAS has not expired. Still no luck. I have tried creating that destination directory both on the NAS before the backup and deleting it so the backup job could create it with the client's credentials but no luck. The usual answer in the Cobian support forums is that there is a permission problem. I agree. But at this point, what can I do to diagnose this further?

    Read the article

  • WSS 3.0 Backup/Restore Root Site Collection to Sub-Site of New Site Collection

    - by bfrancis
    Our intranet was originally setup to be at the root of its site collection. We are trying to change this so that our new internet site will live in the root and the intranet will be a sub-site. At this point I have created a new web application and site collection to house the internet and intranet. I used the 'stsadm -o backup' command to create a backup of our current intranet. I then ran the 'stsadm -o restore' command to restore the intranet site collection to wss/sites/intranet. This seems to have worked as I am able to access the intranet from this location. The issue I now seem to have is that images, sub-sites, etc. are all making reference as if the intranet is still the root site. So for example a link to a sub-site is pointing to wss/department/technology/default.aspx and it needs to point to wss/sites/intranet/department/technology/default.aspx. I am looking for help and/or clarification on two things: 1. Am I approaching the migration of a root site collection to a sub-site the best way? 2. How would I go about updating the link references so that they are based on the intranet now being a sub-site instead of the root site?

    Read the article

  • Windows Image Backup - renamed folder now restore cannot find any backups

    - by Schneider
    A while back I decide to create a couple of Windows Image Backups of my workstation at various points during installation from clean. While doing this I decided to rename the folders containing the VHDs from 'Backup <Date>' to something else of my choosing. I didn't bother testing at the time that the restore still worked. Now I come to use these backups for doing a bare metal restore to a different computer. The problem is restore cannot 'see' any of the backups. So I have deduced that maybe I need to rename them back to the 'Backup <Date>' pattern unfortunately I cannot determine the exact values that would have originally been used here. I have tried by best guest but the images still cannot be found. I have tried doing both a network and a usb hdd restore. No luck on either. P.S. I know I can retrieve files from within the VHDs, the problem is I am trying to save myself time of reinstalling lots of big applications... not trying to recover data.

    Read the article

< Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >