Search Results

Search found 39503 results on 1581 pages for 'database tool'.

Page 297/1581 | < Previous Page | 293 294 295 296 297 298 299 300 301 302 303 304  | Next Page >

  • How to store 250TB of data and develop a backup/recovery plan?

    - by luccio
    I'm really new to this topic, so big apology for stupid questions. I have a school project and I want to know how to store 250TB of data with life-cycle for 18 months. It means every record is stored for 18 months and after this period of time can be deleted. There are 2 issues: store data backup data Due to amount of data I will probably need to combine data tapes and hard drives. I'd like to have "fast" access to 3 month old data, so ~42TB on disk. I really don't know what RAID should I use, or is here any better solution than combining disk and data tapes? Thanks for any advice, article, anything. I'm getting lost.

    Read the article

  • Solr startup script problem

    - by Camran
    I have installed solr and it works finally... I have now problems setting it up to start automatically with a start command. I have followed a tutorial and created a file called solr in the /etc/init.d/solr dir... Here is that file: #!/bin/sh -e # SOLR auto-start # # description: auto-starts solr engine # processname: solr-production # pidfile: /var/run/solr-production.pid NAME="solr" PIDFILE="/var/run/solr-production.pid" LOG_FILE="/var/log/solr-production.log" SOLR_DIR="/etc/jetty" JAVA_OPTIONS="-Xmx1024m -DSTOP.PORT=8079 -DSTOP.KEY=stopkey -jar start.jar" JAVA="/usr/bin/java" start() { echo -n "Starting $NAME... " if [ -f $PIDFILE ]; then echo "is already running!" else cd $SOLR_DIR $JAVA $JAVA_OPTIONS 2> $LOG_FILE & sleep 2 echo `ps -ef | grep -v grep | grep java | awk '{print $2}'` > $PIDFILE echo "(Done)" fi return 0 } stop() { echo -n "Stopping $NAME... " if [ -f $PIDFILE ]; then cd $SOLR_DIR $JAVA $JAVA_OPTIONS --stop sleep 2 rm $PIDFILE echo "(Done)" else echo "can not stop, it is not running!" fi return 0 } case "$1" in start) start ;; stop) stop ;; restart) stop sleep 5 start ;; *) echo "Usage: $0 (start | stop | restart)" exit 1 ;; esac Whenever I do solr -start I get this error: "Error occurred during initialization of VM Could not reserve enough space for object heap" I think this is because of the file above... Also here is where I have solr installed: var/www/solr and here is the start.jar file located: var/www/start.jar Help me out if you know whats causing this. Thanks BTW: OS is ubuntu 9.10

    Read the article

  • Deleting MySQL rows causes lock table error

    - by Dave L
    I had a couple million rows to delete but they can't be deleted at once without this error ERROR 1206 (HY000): The total number of locks exceeds the lock table size So I wrote a script to delete 100,000 rows 10,000 at a time. It ran once but when I run it a second time I get the error on the first attempt to delete 10,000. The way I'm trying to delete the 10,000 rows is to use a delete statement that refers to all 2 million rows but I use a limit clause to affect only 10,000. I've tried adding an "unlock tables;" statement to the script before the first delete but that doesn't help. I still get the lock table error on the first delete. Any ideas how I can do this? Is there a way I can tell it NOT to lock records? I can make sure nothing else is accessing the table.

    Read the article

  • MySQL Server - Got error -1 from storage engine

    - by Bobby
    I am currently trying to restore a MySQL table from the .ibd file. I have been following the instructions on the MySQL reference manual on how to use DISCARD and IMPORT TABLESPACE to replace the .idb files. Discarding the tablespace returns no error and the file is deleted however IMPORTING the replacement .ibd file yields a "Got error -1 from storage engine" error. There doesn't seem to be too much information about what exactly an error -1 is. Does anybody have any further insight as to why an import table space isn't working?

    Read the article

  • MongoDB: ReplicaSet slower than a corresponding Master/Slave config

    - by SecondThought
    Is it true that a mongoDB configured as a replicaset (lets say two nodes + an arbiter) will always be slower than the same DB and server specs but configured as a Master? I've run some tests and found out that for a fresh DB, RS is a little quicker than Master/Slave config but when the DB is getting bigger than ~100k records the latter is getting much snappier. am I missing something here? PS: I was testing it with mongoid driver for ruby.

    Read the article

  • Linux: How to rename old mysqld when upgrading MySQL?

    - by Continuation
    I'm upgrading MySQL from MySQL 5.0 to Percona Server 5.1. I'm planning to just use yum remove and yum install to do the upgrade. However, I read in the documentation that it's a good idea to rename the old mysqld to mysqld-5.0. And if the upgrade doesn't work, I could just revert back to the old version. How exactly does this work? If I use yum remove, doesn't that mean the old mysqld is removed? So how do I rename it? Where is mysqld located? How do I find it? Thanks.

    Read the article

  • SQL Server 2005: Rename DB Server Instance Name?

    - by Code Sherpa
    Hi, Can somebody tell me how to rename the DB server instance name and a DB name in SQL Server 2005? Right Now I Have SERVER/OLDNAME -- oldnameDB I want to change the server instance and also change the db name. I have tried: EXEC sp_renamedb 'oldName', 'newName' and that has changed the dbname as it appers in the tree directory. But, when I do "select @@servername" it is the old name. Also, the MDF and LDF files are still the old name. How do change instance and db names as a clean sweep across the server? Thanks.

    Read the article

  • MySQL asking a user for SUPER privilege to perform a delete.

    - by Fran
    Hello, When trying to do a delete operation on a table, mysql reports the following error: Error code 1227: Access denied; you need the SUPER privilege for this operation. However, my user has this privilege granted for all tables in the schema: GRANT ALL PRIVILEGES ON myschema.* TO 'my_admin'@'%' How come it asks me for SUPER privilege for a delete? Thanks in advance.

    Read the article

  • How do I extract excel data from multiple worksheets and put into one sheet?

    - by user167210
    In a workbook I have 7 sheets(Totals and then Mon to Sat),I want to extract rows which have the word "CHEQ" in its cell (this is a dropdown list with two options-CHEQ/PAID)from all sheets. On my front sheet I used this formula: =IF(ROWS(A$13:A13)>$C$10,"",INDEX(Monday!A$3:A$62,SMALL(IF(Monday[Paid]=$A$10,ROW(Monday[Paid])-ROW(Monday!$I$3)+1),ROWS(A$13:A13)))) This formula works fine for one worksheet (eg. Monday) but is it possible to show the extracted rows from all 6 sheets on the front page? I only have Excel NOT Access. These are the 12 headers on row A12 Col Name Cod House Car Date Discount 2nd Paid Extra Letter Posted The exported data appears like this (this just an example): Col Name Cod House Car Date Discount 2nd Paid Extra Letter Posted 12 Robbs 1244 Ren 11/10 10% 5 CHEQ 0 0 No 15 Jones 7784 Ren 12/10 15% 1 CHEQ 0 0 No 18 Doese 1184 Ren 12/11 12% 1 CHEQ 0 0 No Any ideas on what to do to this formula? I am using Excel 2010.

    Read the article

  • whois returns correct changed date but wrong nameservers

    - by sergio
    I was just told by 1&1 customer support that "the date in the whois could show correctly while the nameservers information being the old one" I though the whole document was stamped at once, not "built" from different data sources. Am I being lied to? or how is that possible? The site is not visible yet, the whois shows correct date the change was made, and the nameservers are still wrong in that report/record/document and of course in the world's dns servers.

    Read the article

  • Purpose of JBoss tables

    - by Rich
    Hi Can anyone point me in the direction of some documentation (or provide the information here) about the following tables, created by JBoss 5.1.0 when it starts up? I know what they are for at a high level, and know why they are there, but I could do with some lower-level documentation about each table's purpose. The tables are: hilosequences timers jbm_counter jbm_dual jbm_id_cache jbm_msg jbm_msg_ref jbm_postoffice jbm_role jbm_tx jbm_user I know that the first two are associated with uuid-key-generator and the EJB Timer Service respectively, while the rest are associated with JBoss Messaging. What I want to know is something along the lines of "jmg_msg stores each message when it is created...", that kind of thing. I wasn't sure where to ask this question, ServerFault or StackOverflow, but I decided it wasn't programming related so thought I should put it here - I hope that's ok! Thanks Rich

    Read the article

  • Do I need to disable access to a publisher database when setting up SQL Server 2000 Transactional Re

    - by Kev
    I have a production database i.e. where there are constant updates and I've configured this to be published to another server using Transactional Replication. When I configure transactional replication I've been doing the following: disable access to the source database backup source DB then restore to subscription server configure replication re-enable DB access to our apps The problem with this approach is scheduling in downtime, having to suspend all the various timed scheduled tasks we run and shutting down access to our various applications that are dependant on this database. Can I just configure transactional replication without disabling access to the publishing database and the subscriber database will correctly catch up? i.e. are all the DML statements queued on the publisher and as soon as the subscriber is ready they are picked off and executed?

    Read the article

  • Is memcache impacting my performence negatively?

    - by iTech
    I am using pressflow 6 and NewRelic seems to suggest that memcache is infact hurting performance as shown below : My settings.php file : # Varnish reverse proxy on localhost $conf['reverse_proxy'] = TRUE; $conf['reverse_proxy_addresses'] = array('127.0.0.1'); # Memcached configuration $conf['cache_inc'] = './sites/all/modules/memcache/memcache.inc'; $conf['memcache_servers'] = array( '127.0.0.1:11211' => 'default', ); ### END Mercury settings written on 2011-11-01T07:12:49-04:00

    Read the article

  • Mirroring MySQL server with diffrent configuration

    - by HTF
    I have to migrate MySQL server to a different data centre so I would like to create another MySQL slave server in new DC and then promote it to a master later on. I previously used LVM snapshots and Percona Xtrabackup for this purpose but this time I've optimized MySQL configuration file that prevents me from using these methods. Old server (backup): innodb_log_file_size = 256M innodb_log_files_in_group = 3 New server (restore): innodb_log_file_size = 512M innodb_log_files_in_group = 2 The Xtrabackup script and LVM snapshots copy the whole directory structure so the MySQL server won't start because there is a different size for InnoDB logs. Is there any solution to avoid a downtime in this case? I can't use mysqldumps as there is around 8000 databases so I would have to take the server down for a couple of hours. I was also thinking to use the old settings with Xtrabackup and then change it once the new server is promoted to a master - less downtime but I'm not sure if this will work? Thank you Regards

    Read the article

  • A complete tool for auditing and archiving emails would be helpful for the community?

    - by Renato Todorov
    Please, don't treat this like a discussion question, I'm looking for direct answers: Yes / No, because... I work on a financial company and I'm needing a tool for email archiving and auditing. The compliance sector is asking for it and I have to provide it. I've searched and found two solutions: MailStore (commercial, Windows only) and Enkive (open source, very poor UI and lack of features). I'm using Postfix as MTA and Courier for IMAP/POP access. I'm almost deciding to write it myself, I have the knowledge needed, but I'm wondering right now if it's worth to put (a lot) more hours to make it open source and user friendly. So my question is, have you ever had the same need? Would this be a helpful solution? Is there any other good tool that I haven't found? Thank you!

    Read the article

  • Sparsity Failure

    - by Lijo
    Hi Team, In the context of data warehouse, could you please explain "Sparsity Failure" of aggregate tables? It would be great if you can explain it with product sales in a store; aggregated by week. It could be easily understood if it is having schema as well as sample data. Thanks Lijo

    Read the article

  • Is Oracle Smartscan really effective?

    - by Vimvq1987
    Oracle is marketing Exadata to our company, and they're focusing on Smartscan, which is, advertised 10x more effective than traditional solution. I don't really understand (and somewhat don't really believe) this solution, because they assume that data is put on separated storage, and no indexes are used, so "1TB data was scanned and only 2MB is returned". So, I want to ask you: is Smartscan really effective. Is it true solution that greatly improves performance or just a buzzword? Thank you so much for your experience

    Read the article

  • Splitting a MySQL DB in two may ease server from "Too many connetions"? I don't think so

    - by Petruza
    I was requested to split a MySQL in two, it's kind of a horizontal partition, in which some rows correspond to one site, and some other correspond to another site. But they want to split it in two DBs in the same MySQL server. I'm no DB expert but I guess keeping them in the same MySQL server with the same amount of memory and processor and the same platform won't improve things. What we're trying to avoid is the "Too many connections" problem.

    Read the article

  • How to find the next generated value for a auto-increment column?

    - by Tim Büthe
    I face some trouble with IBM DB2's auto-increment columns. At first, all my columns were defined as GENERATED ALWAYS, but since I had trouble with this when using the "db2 import ..." command, I changed them to GENERATED BY DEFAULT. This is necessary, sinceI need the IDs to be consistent, because other tables reference them. So using "db2 import ... modified by identityignore ..." isn't an option. When I now import data, the IDs are inserted correctly, but everytime I do this, I have to remember to set a new start for the auto-increment column by getting the highest Id+1 and alter the column like this: SELECT MAX(mycolumn)+ 1 FROM mytable; ALTER TABLE mytable ALTER COLUMN mycolumn RESTART WITH <above_result>; If I forget this, an Insert-Statement will fail with an duplicate PK error, since the auto-increment column is the primary key. So my question is: Is there a way to find the next value for an auto-increment column, so I could write Statements that would check, if this value is less then the SELECT MAX and needs to be set? Or: Isn't this whole thing as complicated as it seems to me? Could I somehow import data, preserving the IDs and have the auto-increment column still working as expected?

    Read the article

  • Tool for creating system image from DOS (should be USB bootable) ..?

    - by Ahmad
    I want to know if there's any decent tool which can be used to create the system image from DOS .. What I specifically want to do is to put the program in my FAT32 formatted USB, then boot the target computer from the USB so that the tool runs, and then it should be able to create a complete system image of the entire system, and store it on the USB itself .. Please note that I can ONLY do this from boot time because of other limitations .. I cannot go into any OS to do it from there .. So I need tools which can do this at boot time from DOS ..

    Read the article

  • how to warehouse data that is not needed from sql server

    - by I__
    I have been asked to truncate a large table in sql server 2008. The data is not needed but might be needed once every two years. It will NEVER have to be changed, only viewed. The question is, since I don't need the data on a day-to-day basis, what do I do with it to protect and back it up? Please keep in mind that I will need to have it accessible maybe once every two years, and it is FINE for us if the recovery process takes a few hours. The entire table is about 3 million rows and I need to truncate it to about 1 million rows.

    Read the article

< Previous Page | 293 294 295 296 297 298 299 300 301 302 303 304  | Next Page >