Search Results

Search found 5457 results on 219 pages for 'necessary'.

Page 68/219 | < Previous Page | 64 65 66 67 68 69 70 71 72 73 74 75  | Next Page >

  • Software to manage "application profiles", which services are running, etc., on Windows?

    - by Lasse V. Karlsen
    I am looking for a particular type of program. I use my computers in various "modes", like gaming, programming, just plain surfing, etc. What I'd like to find is a program that can help me manage these modes. For instance, while programming I might use SQL Server, but while gaming I don't want those services running, but perhaps I'd like Steam to run instead. Basically, the program type I'm looking for is a visual program that allow me to quickly switch modes, and when I do, the program would start and stop the necessary services and applications in order to leave one mode and enter another. I've looked at the programs related to startup management, and I haven't found one that lets me do what I want. At the moment I have batch files, but they're not very good at conveying problems or other things, I'd like a more visual program.

    Read the article

  • Converting gzip files to bzip2 efficiently

    - by sundar
    I have a bunch of gzip files that I have to convert to bzip2 every now and then. Currently, I'm using a shell script that simply 'gunzip's each file and then 'bzip2's it. Though this works, it takes a lot of time to complete. Is it possible to make this process more efficient? I'm ready to take a dive and look into gunzip and bzip2's source codes if necessary, but I just want to be sure of the payoff. Is there any hope of improving the efficiency of the process?

    Read the article

  • Mysql table Crashed during Optimize, table repair taking too long

    - by hellohellosharp
    One of my vital tables was being optimzied and MySQL crashed in the middle of it. It then said the table was corrupt. I am running repair but it is taking over an hour, I need this table up ASAP (I will even truncate it if necessary). Please help me get this solved. The table has about 54 million rows. This is a MyISAM. Any additional information needed, just ask. Here are the contents of my my.cnf: [mysqld] max_connections = 850 max_user_connections = 850 query_cache_size = 128M skip-external-locking key_buffer_size = 64M max_allowed_packet = 8M table_open_cache = 256 sort_buffer_size = 4M net_buffer_length = 16K read_buffer_size = 1M read_rnd_buffer_size = 1M myisam_sort_buffer_size = 32M innodb_file_per_table tmp_table_size = 100M max_heap_table_size = 64M thread_cache_size = 8 wait_timeout=25 interactive_timeout=25 table_cache=600 innodb_buffer_pool_size = 4G innodb_thread_concurrency = 8 innodb_flush_method = O_DIRECT # This setting allows the use of asynchronous I/O in InnoDB. # The following files track usage of this resource: # - /proc/sys/fs/aio-max-nr # - /proc/sys/fs/aio-nr # Default limit is 65536, of which a single instance of mysql uses 2661 out of the box innodb_use_native_aio = 1

    Read the article

  • Processor speeds on my machine don't live up to manufacturer hype

    - by atch
    Why am I not seeing the promised speed claims of processor manufacturers on my computer? Producers of processors claim that their product can perform so many thousands (or millions) of operations per second. And yet on my machine (4GB, 3500hz), the typical program (Word, Visual Studio etc.) takes at least 10 seconds to start. I've formatted my hard drive and ticked all the necessary boxes to optimize my machine and yet I'm not seeing the promised speeds. Say it takes Outlook ten seconds to load. How many millions of operations does it really go through in order to start up?

    Read the article

  • What is the best package repository for Debian ?

    - by Dan Sosedoff
    Hello everybody I have a question regarding packages repository in Debian (Lenny). A new server at rackspace cloud service comes with nothing (plain system) so i had to upgrade system and install all necessary packages by myself. But the problem is that default repositories does not provide newer versions of software i need (PHP is 5.2.6, comparing to 5.2.13). In CentOS i used to get all new packages from third-party sources, which works really well till now. I have talked to guys from rackspace about it and looks like they cant provide anything except kernel headers. And i dont really want to build from source. So, whats the way to get newer software ? Any good third party sources? Thanks.

    Read the article

  • Easy Linux distribution with newer packages than Ubuntu?

    - by sweetiecakes
    I'm a programmer and sysadmin, and I'm looking for a better Linux distro to use than Ubuntu. It certainly is a well-polished, nice distribution to use, but a lot of the programs available in the Ubuntu repositories are very old versions. Installing PPA's or compiling from source just isn't very nice. I'd love to use something like Arch Linux that uses a rolling release cycle, but I really don't want to configure my system from scratch, and support for ATI Catalyst drivers is necessary. I just want to pop the CD in, install and start using my computer - just like with Ubuntu. Additionally, if you know of a package like ubuntu-restricted-extras for the distribution, that'd be nice! What do you suggest?

    Read the article

  • Flexible classroom environments (OS, Office)

    - by HannesFostie
    I work in the IT department of a training center, we still offer XP and Office 2003 trainings but also offer Vista and Win7 and Office 2007. Currently, we use VMs on VMware Server but this is obviously not a superb choice. We're thinking of implementing something like VDI (brainstorm phase, we hardly have any details) but I decided to check here if people would have some clever alternatives. Requirements: * Flexible when it comes to deployment * Centralized management would be a big plus * Allow for different software, whether they be compatible or not (all of office except for outlook can be installed simultaneously. for outlook you need to choose between 2003 or 2007) * Allow for different OS We have a big enough budget to implement a proper SAN environment to accomodate the virtualization of the solution, whatever kind it may be. A support contract will probably be necessary as well, because we need to be able to offer quick solutions to problems and with only 2 sysadmins that is simply impossible to guarantee.

    Read the article

  • Downtime scheduling - can it be automated?

    - by Nnn
    When servers need to have scheduled downtime at my workplace we follow a process roughly like the following: Propose time for work to take place on specific box/s Lookup list of stakeholders for specific box/s Seek approval from stakeholders (service owners/management etc) via email Incorporate changes to proposed time if necessary, repeat step 2 until.. Now everyone is happy with the time, send out a notification via email of the time, ask Staff who care about when the box is going down manually add it to their calendars some stakeholders the staff doing the work Do the actual work Is there an OSS project that we could use to automate this process? My googling has been fruitless so far. Will we need to build something ourselves? Would anyone else be interested in something like this?

    Read the article

  • Problems to export java home and to find or create .bashrc in Mac OS X 10.6

    - by casiopea
    Hello, I need to install a program for my studies and, this program, need java to run. When I try to perform the installation say that cannot find the JDK; since the JDK is already installed by default in mac, the problem is export the java home. I´ve try a lot and I cant do it! I know that I have to add a line in a .bashrc file (or .profile, or .bash_profile) I´ve created all those files, at different times but nothing... I´m a new mac user, but I use Linux too and I dont know what happened, I just need to export java home to perform my work... and is really necessary for me to add environment variables too. Thanks for your help

    Read the article

  • Unmounting a zfs pool while it is shared with sharenfs

    - by Ted W.
    I have a Solaris (open indiana) system which is getting poor disk write performance. In order to enable ZIL in this version of zfs I need to add a line to /etc/system. This will not take affect until I've unmounted and remounted the zpool. The trick is that this spool is shared via nfs to about 200 other servers to host users' home directories. I can guarantee that no users will be accessing the disks during this period of maintenance but I would like to avoid having to issue an unmount for 200 systems in order to unmount the disk on the Solaris box. My question is, with sharenfs, is it necessary to have all systems disconnected before unmounting the filesystem on the host? If it's possible, how do you go about it? I've tried unmounting already, the normal way, and it reports the disk is busy. There is no lsof in Solaris and pfiles (I think that's what it was) does not show anything obviously using the mounts.

    Read the article

  • Steps to install Windows 7 64bit on RAID 0 (striping)?

    - by marco.ragogna
    I will receive in some days 2x500 GB hard disks (ST3500418AS) and an ASRock 890GX Extreme 3. My idea is to install onto it Windows 7 64-bit in RAID0 configuration (striping). I wondering which steps should I follow, due to the fact I never did it before. Should I install Windows 7 on a single disk and apply the RAID0 later, or should I perform some step through BIOS first and install then Windows 7? If you can, please list me all necessary steps I should follow. Thank you in advance, Marco

    Read the article

  • Ruby, Rails & MySQL parity between Mac Client (10.6) & XServe (10.5)

    - by Meltemi
    We're setting up a RoR setup with Development on Mac OS X Client (10.6.3) and then using a Mac OS X Server (10.5.8) for testing and eventually deployment. I'd like to get as many systems in sync on these machines as possible. Wondering if there are any pitfalls. I seem to understand what's necessary under Client but Server has some hardwired stuff that I want to make sure doesn't break...or is updated correctly. Currently installed on both machines we have: OS X Client (10.6.3): Ruby 1.8.7 Rails 2.3.5 MySQL (not installed yet) OS X Server (10.5.8): Ruby 1.8.6 Rails 2.3.5 MySQL Ver 14.12 Distrib 5.0.82 Any suggestions...Ideally from someone who's done this on Leopard Server as well but I'll listen to general tips & proceedures

    Read the article

  • handling long running large transactions with perl dbi

    - by 1stdayonthejob
    I've got a large transaction comprising of getting lots of data from database A, do some manipulations with this data, then inserting the manipulated data into database B. I've only got permissions to select in database A but I can create tables and insert/update etc in database B. The manipulation and insertion part is written in perl and already in use for loading data into database B from other data sources, so all that's required is to get the necessary data from database A and using it to initialize the perl classes. How can I go about doing this so I can easily track back and pick up from where the error happened if any error occurs during the manipulation or insertion procedures (database disconnection, problems with class initialization because of invalid values, hard disk failure etc...)? Doing the transaction in one go doesn't seem like a good option because the amount data from database A means it would take at least a day or 2 for data manipulation and insertion into database B. The data from database A can be grouped into around 1000 groups using unique keys, with each key containing 1000s of rows each. One way I thought I could do is to write a script that does commits per group, meaning I've got to track which group has already been inserted into database B. The only way I can think of to track the progress of which groups have been processed or not is either in a log file or in a table in database B. A second way I thought could work is to dump all the necessary fields needed for loading the classes for manipulation and insertion into a flatfile, read the file to initialize the classes and insert into database B. This also means that I got to do some logging, but should narrow it down to the exact row in the flatfile if any error occurs. The script will look something like this: use strict; use warnings; use DBI; #connect to database A my $dbh = DBI->connect('dbi:oracle:my_db', $user, $password, { RaiseError => 1, AutoCommit => 0 }); #statement to get data based on group unique key my $sth = $dbh->prepare($my_sql); my @groups; #I have a list of this already open my $fh, '>>', 'my_logfile' or die "can't open logfile $!"; eval { foreach my $g (@groups){ #subroutine to check if group has already been processed, either from log file or from database table next if is_processed($g); $sth->execute($g); my $data = $sth->fetchall_arrayref; #manipulate $data, then use it to load perl classes for insertion into database B #. #. #. } print $fh "$g\n"; }; if ($@){ $dbh->rollback; die "something wrong...rollback"; } So if any errors do occur, I can just run this script again and it should skip the groups or rows that have been processed and continue. Both these methods is just variations on the same theme, and both require going back to where I've been tracking my progress (in table or file), skip the ones that've been commited to database B and process the remaining data. I'm sure there's a better way of doing this but am struggling to think of other solutions. Is there another way of handling large transactions between databases that require data manipulation between getting data out from one and inserting into another? The process doesn't need to be all in Perl, as long as I can reuse the perl classes for manipulating and inserting the data into the database.

    Read the article

  • Developer Laptop with SQL Server 2008 can't login to SSIS when offsite

    - by wizlb
    When I bring my Windows XP (SP3) laptop home I can still login as my domain account because Windows caches the info necessary to authenticate me when the domain controller isn't around. However, when I try to connect to Integration Services from within SQL Server Management Studio, it generates SSPI context errors. The only way it works is if I connect to the office with VPN or if I'm at the office where the domain controller is. I have both SQL Server Agent and SQL Server Integration Services 10 running under local computer accounts. It seems that the only option to connect to Integration Services from within Management Studio is to use Window authentication. Is there any way to do this when I'm not connected to the office? Why don't these services use the cached info just like Windows Login? Thanks.

    Read the article

  • Working on a jQuery plugin : viewport simulator

    - by pixelboy
    Hi community, i'm working on a jQuery plugin that's aiming to simulate one axis camera movements. The first step to achieving this is simple : use conventional, clean markup, and get the prototype to work. Heres how to initiate the plugin : $('#wrapper').fluidGrids({exclude: '.exclude'}); Here's a working demo of the WIP thing : http://sandbox.test.everestconseil.com/protoCitroen/home.html Where am I having issues : Is there a fast way to detect if parent of each target (clickable animated element) is a href link, and if so to save this url ? My clone uses the original background image to then animate it.fade it to black/white. But when you clik on an element, image url is found, but doesn't seem to be injected. Do you see anything ? Finally, about animation of elements : as you can see in source code, I'm using the container $('#wrapper') to position all animated children. What would be the perfect properties to apply to make this cross browser proof ? Here's the source code for the plugin, fully commented. (function($){ $.fn.extend({ //plugin name - fluidGrids fluidGrids: function(options) { //List and default values for available options var defaults = { //The target that we're going to use to handle click event hitTarget: '.animateMe', exclude: '.exclude', animateSpeed: 1000 }; var options = $.extend(defaults, options); return this.each(function() { var o = options; //Assign current element to variable var obj = $(this); //We assign default width height and positions to each object in order to get them back when necessary var objPosition = obj.position(); //Get all ready to animate targets in innerViewport var items = $(o.hitTarget, obj); //Final coords of innerViewport var innerViewport = new Object(); innerViewport.top = parseInt(objPosition.top); innerViewport.left = parseInt(objPosition.left); innerViewport.bottom = obj.height(); innerViewport.right = obj.width(); items.each(function(e){ //Assign a pointer cursor to each clickable element $(this).css("cursor","pointer"); //To load distant url at last, we look for it in Title Attribute if ($(this).attr('title').length){ var targetUrl = $(this).attr('title'); } //We assign default width height and positions to each object in order to get them back when necessary var itemPosition = $(this).position(); var itemTop = itemPosition.top; var itemLeft = itemPosition.left; var itemWidth = $(this).width(); var itemHeight = $(this).height(); //Both the original and it's animated clone var original = $(this); //To give the if (original.css('background-image')){ var urlImageOriginal = original.css('background-image').replace(/^url|[("")]/g, ''); var imageToInsert = "<img src="+urlImageOriginal+"/>" } var clone = $(this).clone(); original.click(function() { $(clone).append(imageToInsert); $(clone).attr("id","clone"); $(clone).attr('top',itemTop); $(clone).attr('left',itemLeft); $(clone).css("position","absolute"); $(clone).insertAfter(this); $(this).hide(); $(clone).animate({ top: innerViewport.top, left: innerViewport.left, width: innerViewport.bottom, height: innerViewport.right }, obj.animateSpeed); $("*",obj).not("#clone, #clone * , a , "+ o.exclude).fadeOut('fast'); //Si l'objet du click est un lien return false; }); }); }); } }); })(jQuery);

    Read the article

  • Extract data from specific range of cells in multiple worksheet in multiple files.

    - by Michele
    Extract data from specific range of cells(always the same cells) in multiple worksheet in multiple files. 1 file=1 day. I have 6 technicians each day of the week, Monday thru Friday. So, 5 files with 6 worksheets. I have entered specific info in specific cells of every work sheet. The range is constant(the same address in EVERY worksheet in every file.) So, I need a formula to extract and calculate the data in the given range and dump it into another spreadsheet. I can forward an example a file if it will help anyone to answer my question. Or more explanation if necessary is available upon request. JUST PLEASE SOMEBODY HELP ME!!!!! Thank you all in advance. Regards, Michele

    Read the article

  • Applescript to connect to bluetooth device

    - by Mark117
    I'm trying to create an applescript to allow me to connect to a bluetooth device by it's Bluetooth ID. So far, i've managed to get an applescript to turn on bluetooth if it's off. Here's the code: # This is only necessary, if AppleScripts are not yet allowed to change checkboxes tell application "System Events" to set UI elements enabled to true # Now change the bluetooth status tell application "System Preferences" set current pane to pane id "com.apple.preferences.bluetooth" tell application "System Events" tell process "System Preferences" # Enabled is checkbox number 2 if value of checkbox 2 of window "Bluetooth" is 0 then click checkbox 2 of window "Bluetooth" end if end tell end tell quit end tell Would someone know if and how it's possible to set up a new bluetooth device and if it'd be posible to connect to a device based on it's device name/ it's device bluetooth ID? I've also tried to record the action in Automator but for the "set up new device" option, Automator just tells me: "click on "" button". Thanks

    Read the article

  • Drive configuration for 5 large databases

    - by Mr. Flibble
    I've got 5 databases, each 300GB, currently on a RAID 5 array consisting of 5 drives. All the databases are used heavily, at the same time, so drive speed is an issue. Would I see better performance if I got rid of the RAID 5 configuration and just put each database on a separate drive? The redundancy provided by RAID 5 is not necessary due to mirroring elsewhere. Will the server then be able to perform reads / writes to different databases drives in parallel? More so at least than when it's in RAID? This is all on Windows 2003 / SQL 2008.

    Read the article

  • Where can I collaborate with my friend on source code in real time?

    - by Carson Myers
    I mean, other than a conference room :) Using google docs, I can upload any kind of file and view it with other people, watch them edit it in real time, with a live chat happening in the same window. This is awesome. How can I do the same thing with source code? I'm looking for a web application where I can upload source files that will be displayed in some kind of editor, with syntax highlighting, and allow others to view it and edit it in real time. Preferably with a live chat also, but not necessary. Does anybody know where I can find this?

    Read the article

  • Is there any way to use the xscreensaver gltext to monitor system information?

    - by Shane
    I am trying to get the X screensaver gltext to monitor my system temp and maybe some other stats. Writing a script that puts together the stats periodically is no problem, but the main thing I'm running into is that gltext doesn't refresh - whatever text I feed it stays there. So, for example, I run this command: $ /usr/lib/xscreensaver/gltext -text "`cat /proc/acpi/thermal_zone/THRM/temperature`" and get a gltext window showing: temperature: 60 C I can manipulate it and format it as necessary, but as my CPU heats up and cools down, it doesn't update, even if I include time variables that do. I have a script that feeds gltext the time as the first line and the temp as the second line - and although the time updates continuously as the time changes, the temperature value remains the same as whenever the screensaver started. Is it possible to do what I want, which is change the text every 60 seconds if the temperature changes?

    Read the article

  • How to install imagemagick on windows 7

    - by learner
    How to install image magic at windows 7. I followed these instruction To install IMagick on windows xp (php 5.2.x) 1.) download and install ImageMagick-6.5.8-7 Q16-windows-dll.exe imagemagick.org/download/binaries/ ImageMagick-6.5.8-7-Q16-windows-dll.exe 2.) download php_imagick_dyn-Q16.dll from: valokuva.org/outside-blog-content/ imagick-windows-builds/080709/ copy dll to [PHP]/extension dir and rename it to php_imagick.dll 3.) You have to edit your php.ini file and add new extension "extension=php_imagick.dll" 4.) Save ini file and restart apache server. (If necessary, restart your windows) 5.) phpinfo() should show imagick enabled. after that I execute a sample script but its not working. it shows the Imagic class missing error. Please help me to install Imagick. :-(

    Read the article

  • keeping URL domain the same when pointing A record to a hosting account

    - by kwight
    Hello, I have a new WordPress website and a legacy billing system. For technical reasons, they cannot be on the same hosting plan. The hosting account for billing (and the original abc.com website) also manages DNS and mail. I'm trying to incorporate the new website under the same domain, eg. abc.com (website, on a different hosting account) and billing.abc.com (billing). I assume the answer is having a different A record for abc.com. I currently have a CPanel shared hosting account to use for the website (but can upgrade if necessary). How would I set this up in CPanel, so that the URLs work properly? Do I need a dedicated IP and then add the domain as an add-on domain? Thanks

    Read the article

  • Howto manage website and user permissions with apache on Mac os X

    - by Sander Versluys
    I have lots of different websites in a Development directory in my home dir. While developing, files get saved under my username but websites configured under apache it's permission need to be set as _www user and group. What's the best way to handle this? Do I run apache under a different user/group? Do I run my development tools under a different user? Do add myself to the _www group? (seems like it doesn't work btw) I've just switched to a mac and I'm trying to find a smooth development workflow, so it would be best if i could just run the necessary tools, save some files and be able to test the website without much hassle. Thanks!

    Read the article

  • Send an Email at a future date

    - by Ryan
    I'd like to write up an email that gets sent out in a few days. I'd prefer to use Gmail, but I could use some other client if necessary. It doesn't look like Gmail has this feature in their labs anywhere, but it could just be hiding somewhere. Any ideas? EDIT: a bit more backstory on my particular situation. My wife is out of town for three weeks, and I've decided to email her every day while she's out. Unfortunately, I myself am going camping this weekend, so I wanted to pre-record a message that gets sent while I'm out. Unfortunately, FutureMail and FutureMe both are for sending email to yourself, probably for anti-spam reasons. I guess the best solution is to use thunderbird on my laptop (so it's shielded from power outages). Seems a little excessive to keep a computer running just to send a few emails, but whatever gets the job done :).

    Read the article

  • Intentionally hogging Wi-Fi bandwidth?

    - by endolith
    I've noticed that Wi-Fi signals are interfering with a product I'm developing, and I'd like to generate as much Wi-Fi noise as possible for testing purposes. Is there any better solution than, say, dragging large files from one computer to another? Ideally I'd like one computer to just generate a stream of data ex nihilo and stream it to the other computer where it will just be obliterated, so it hogs bandwidth without reading or writing the hard drives. I'm in Windows, though, so there's no /dev/random or /dev/null. And it would be cool if I could vary the bandwidth, too, but not necessary.

    Read the article

< Previous Page | 64 65 66 67 68 69 70 71 72 73 74 75  | Next Page >