Search Results

Search found 3618 results on 145 pages for 'huge'.

Page 56/145 | < Previous Page | 52 53 54 55 56 57 58 59 60 61 62 63  | Next Page >

  • I need a script to lockdown the system time to users via gpedit.msc

    - by Chester
    I need to lockdown the system time on a number of PCs via gpedit.msc and then removing administrators from the group and then adding 'administrator' and 'polling'. Can I do this via a script? Essentially I have to; Run gpedit.msc Computer Configuration Windows Settings Security Settings Local Policies User Rights Assignment Double Click Change the system time Select Administrators Click Remove Click Add User or Group Type Administrator Click Check Names Type polling Click Check Names OK Apply OK Logoff I have to do this for a huge number of computers so is there a batch file I could run on each PC to do this? Your help would be very much appreciated. Best Regards

    Read the article

  • How to keep program always in memory (no swapping)

    - by Yossarian
    I'm using KeePass (on Windows 7) for storing passwords. The application is running on my laptop almost 100% of the time, but I'm using it rarely. The problem is, that activating the window after long time takes huge amount of time, I'd say that OS realized the application is no longer used and swapped it. I don't want OS to swap my KeePass (because of the loading lag, and also because of security - I don't like the idea of password keeping program's memory to be stored anywhere on HDD). Is there any possibility how to tell OS not to swap some program? Or, is this achievable by some programming?

    Read the article

  • How to "demote" all titles and headings in Word 2010?

    - by dangowans
    I built a large help document for an application I wrote. I used all the default styles in Word 2010, including "Title", "Heading 1", "Heading 2", etc. Sadly, when I generated the Table of Contents, Titles were not included. I'm also now using chmProcessor to automatically generate a website from the document, and it's not including Titles in its Table of Contents either. I'd like to make all Titles into Heading 1s, all Heading 1s into Heading 2s, and Heading 2s into Heading 3s, etc. Is this possible without a huge manual effort? (I'm sure there's a better word than "demote" for this.)

    Read the article

  • Incrementing ticket numbers each time I print

    - by Danny
    I have an excel sheet where I have a set 4 identical tickets to print per page which we use for stock takes. Rather then creating a huge document with 1000 pages for 4000 tickets each with their own unique ticket number (starting from 1) I would like to find a Macro or function which will print a page with 4 tickets on (1,2,3,4) then continue to print another with (5,6,7,8) and so on. I have found some code that people have already written but it has only applied to one number changing per page rather than 4 simultaneously and being a complete visual basic novice, I was unable to change the code to suit my preferences. If someone could explain simply how I could achieve this I would be very very grateful :)

    Read the article

  • Upgrade from "audit mode" (pre-cloning) Windows 8 to Windows 8.1?

    - by Display Name
    I have a Windows 8 in audit mode with a lot of applications installed, custom configurations done, and an answer file prepared, ready to be sysprepped for cloning. How do I upgrade to Windows 8.1, when I can't go into the store (Metro apps don't work in audit mode)? If I run sysprep then create a normal account so I can get the upgrade from the store, I suppose there's no way to go back to audit mode, and that's a huge problem as I want to retain the particular account settings I have configured for the audit mode account as a default account. What do I do??

    Read the article

  • Is there any way to do "mail server parking"?

    - by percyboy
    I am managing a mail server, which will be temporarily closed for three or four days due to the data center maintenance. I want to find a solution to (completely or partly) solve the lost mails during this unavailable period. Because the data volume is huge, it is very hard to migrate it to other data center. One approach I think out is to setup a temporary mail server in other data center, and when new mail received, the mail server automatically sends a return mail to tell the sender "We are temporarily closed for three or four days. Please send the mail later or contact in other means." I am wondering is this approach possible with existed mail server ? Or something better available ? (free solution is preferred for it is only for temporary)

    Read the article

  • Dangers of the pyton eval() statement

    - by LukeP
    I am creating a game. Specifically it is a pokemon battle simulator. I have an sqlite database of moves in which a row looks something like: name | type | Power | Accuracy | PP | Description However, there are some special moves. For said special moves, their damage (and other attributes not shown above, like status effects) may be dependant on certian factors. Rather than create a huge if/else in one of my classes covering the formulas for every one of these moves. I'd rather include another column in the DB that contains a formula in string form, like 'self.health/2'(simplified example). I could then just plug that into eval. I always see people saying to stay away from eval, but from what I can tell, this would be considered an acceptable use, as the dangers of eval only come into play when accepting user input. Am I correct in this assumption, or is there somthing i'm not seeing.

    Read the article

  • Looking for a way execute a task on all files in a directory (recursively) on Windows

    - by stzzz1
    I have a huge number of mp4 video files that needs to have a volume boost. I need a way to execute a ffmpeg audio filter on all files in a specified base directory (and in subdirectories as well). My problem is that I'm working on a Windows computer and I have no knowledge of its shell syntax. I would like to do the equivalent of what this bash script does : TARGET_FILES=$(find /path/to/dir -type f -name *.mp4) for f in $TARGET_FILES do ffmpeg -i $f -af 'volume=4.0' output.$f done I spent quite some time this afternoon looking for a solution but the recursive nature of what I need (that is so simple with find!) isn't too clear. Any help would be greatly appreciated!

    Read the article

  • How to embed word doc as background picture of an Access report using .EMF or equivalent ?

    - by iDevlop
    My company's standard paper has logo, address and all the details in the right margin with a vertical blue line. I have that as a word template. I want to have the same thing as the background of my Invoices report. I managed to do that 5 years ago by saving to EMF format (vector format, prints out nicely) and putting the file as the background of the report. Now my company is moving, and I need to change the address on my invoices, but I can't find out how I did to convert the word doc to EMF. Any suggestion ? By EMF or another process, but I want to avoid BMP, which is huge and does not print nicely. Thanks !

    Read the article

  • can I create disk partition for dual-boot Ubuntu on Windows 7 machine without Windows reinstall?

    - by EndangeringSpecies
    I want to setup dual boot Ubuntu on my machine in a separate partition. Plus, ideally, I want to get another, 3rd, partition for further OS experimentation. The hard drive is huge, hundreds of gigs, and essentially unfilled. The machine runs Windows 7 Home. Online I have seen mention of creation of partitions from inside Windows 7. But, I have also heard claims that to create the partition to house Ubuntu Windows has to be reinstalled, frying all the data on the machine. So, which one of these claims are right? Can you create additional partitions for other OS on a big Windows 7 hard drive without reinstall?

    Read the article

  • Stream tar.gz file from FTP server

    - by linker
    Here is the situation: I have a tar.gz file on a FTP server which can contain an arbitrary number of files. Now what I'm trying to accomplish is have this file streamed and uploaded to HDFS through a Hadoop job. The fact that it's Hadoop is not important, in the end what I need to do is write some shell script that would take this file form ftp with wget and write the output to a stream. The reason why I really need to use streams is that there will be a large number of these files, and each file will be huge. It's fairly easy to do if I have a gzipped file and I'm doing something like this: wget -O - "ftp://${user}:${pass}@${host}/$file" | zcat But I'm not even sure if this is possible for a tar.gz file, especially since there are mutliple files in the archive. I'm a bit confused on what direction to take for this, any help would be greatly appreciated.

    Read the article

  • Show Excel column filter information in cells

    - by Alex
    We have a sheet with a huge number of columns and filtering is often used to navigate to the correct data. The problem is that sometimes its not obvious that the filter has been applied , the visual cue is very subtle. Is it possible to show some data via a formula or VBA about the filter inside another cell? Something like this: Just knowing if the filter is active would be a good help, knowing what columns have active filters applied to them would be icing on the cake. Ideally they update automatically. I dont have ownership of the spreadsheet so cant make major changes to its structure or anything but VBA is fine. Any ideas?

    Read the article

  • Sharepoint 2007 reset permission inheritance

    - by e-mre
    I have this SharePoint 2007 document library which has several levels of folders and files. Some folders in the middle of the hierarchy do not inherit permissions from their parents and have their unique permissions defined. It is a huge library and there are many folders like this. I am currently changing the permission model of my library and I want to reset all those unique permissions and have all of them inherit permissions from the library root. (Something like "Replace child object permissions" checkbox available in windows files system security window) If this is not possible, seeing a list of folders that have their unique permissions defined would also do.

    Read the article

  • Sorting Files into Subfolders based on EXIF Date

    - by honestor
    I have a huge directory from a HDD recovery that contains 70000+ JPEG files. I tried playing around with some AppleScripts, that I found, but had no luck. I already installed EXIFtool, which might be useful for this task. The current directory structure is as follows: dir001 - file0001.jpg ... - file9999.jpg dir002 - file0001.jpg ... - file9999.jpg ... dir070 - file0001.jpg - ... - file9999.jpg The files mostly have EXIF Data, but sometimes there are Files without metadata. Now I hope to be able to sort and rename these files into folders based on the date: 1999 - 1999 01 31 - 1999_01_31_-_22_59_59.jpg 2000 - 2000 05 20 - 2000_05_20_-_21_59_59.jpg - 2000_05_20_-_22_59_59.jpg I figured Applescript/Automator might come in handy for this, however every other solution would be welcome, too!

    Read the article

  • SSH and Latent Connections (e.g., satellite connections)

    - by user71494
    Most of the week I live in the city where I have a typical broadband connection, but most weekends I'm out of town and only have access to a satellite connection. Trying to work over SSH on a satellite connection, while possible, is hardly desirable due to the high latency ( 1 second). My question is this: Is there any software that will do something like buffering keystrokes on my local machine before they're sent over SSH to help make the lag on individual keystrokes a little bit more transparent? Essentially I'm looking for something that would reduce the effects of the high latency for everything except for commands (e.g., opening files, changing to a new directory, etc.). I've already discovered that vim can open remote files locally and rewrite them remotely, but, while this is a huge help, it is not quite what I'm looking for since it only works when editing files, and requires opening a connection every time a read/write occurs. (For anyone who may not know how to do this and is curious, just use this command: 'vim scp://host/file/path/here)

    Read the article

  • What's the best way to be able to reimage windows computers?

    - by mos
    I've got a low-end machine for testing our software. It needs to be tested under various versions of Windows, so I was planning installing each one on its own partition. Then I realized that after testing our software, I'd want to roll back to the previous, clean state. I don't want to use any virtualization software because it tends to interfere with the workings of our app. That said, what's the best way to achieve my goal? Norton Ghost? Edit: I work for a pretty monstrously huge organization. Money is no object here (and sometimes, if the wrong people get wind of it, "open source" software is bad).

    Read the article

  • Laptops and Visual Studio 2010

    - by Chad
    I have three days to get a new laptop before I need one that can run Visual Studio 2010. Normally I have two instances open, an instance of SQL Management studio, IIS, and a few web browsers. Who develops regularly on one, what are the specs I should look at targeting? I haven't bought a laptop in 5 years, so I'm way out of touch with where they are at as far as specs go. Thanks. ps. Budget isn't a huge concern, but I'm trying to keep it reasonable.

    Read the article

  • Why is an Ext4 disk check so much faster than NTFS?

    - by Brendan Long
    I had a situation today where I restarted my computer and it said I needed to check the disk for consistancy. About 10 minutes later (at "1%" complete), I gave up and decided to let it run when I go home. For comparison, my home computer uses Ext4 for all of the partitions, and the disk checks (which run around once week) only take a couple seconds. I remember reading that having fast disk checks was a priority, but I don't know how they could do that. So, how does Ext4 do disk checks so fast? Is there some huge breakthrough in doing this after NTFS came out (~10 years ago)? Note: The NTFS disk is ~300 GB and the Ext4 disk is ~500 GB. Both are about half full.

    Read the article

  • Something keeps deleting my downloaded files

    - by corroded
    I have been using utorrent for years now and recently I was surprised that I had 24GB free. I thought that was because I deleted some unused apps, but after awhile, I noticed my Torrents folder was gone(I put finished torrents in my Downloads/Torrents folder) I thought I accidentally deleted it(I use rm -r to delete huge files) so I shrugged and tried to download those 24GB back(after banging my head for the sheer stupidity) This morning, I noticed that again, my Torrent folder was gone! This made me think that something MUST be deleting my torrent files. I am not sure but my hunch is uTorrent(so I just upgraded it) or something else entirely. This is getting frustrating, so I hope someone can help me on this. My only guess is when I do CMD + w (I'm on a Mac, OSX Lion), it closes the window and somehow deletes the torrents? I am downloading files again now and will try to document what I do the tomorrow so I can add more input here.

    Read the article

  • Where to look for regular scripts?

    - by fontan
    It seems to me that our server freezes every 30 days around noon due to the huge utilization of xvda data transfer partition - writes are 50 times higher than normally (according to the health monitor in plesk). This seems to me as the reason why the apache & co becomes instable as (for example) all apache's processes are waiting to write their log (according to the service's full status). I am, however, unable to find any scheduled task that would be executed during that period. I have checked both cron and anacron setup and there is only one monthly anacron task which is not executed (according to the /var/log/cron - and there is nothing unusual) around noon. Are there any other places where to look for periodical processes? (I am just about to ask server's provider the same question about any external maintenance run around this time but I don't expect them to run anything time/resource consuming during the day.)

    Read the article

  • What is the significance of these different width, height and resolution parameters?

    - by ??????? ???????????
    An image with a pixel resolution of 640 x 480 has additional dimension and resolution parameters according to exiftool. I'm unsure what they mean. Why are the X / Y Resolution parameters the same?Should they not reflect the pixel dimensions of the image? What does Exif Image Size mean and how is it different from the pixel dimensions? What is the focal plane? Does it have any relation to the device used to capture this image? $ exiftool evil1.jpg | egrep 'Width|Height|Resolution' X Resolution : 180 Y Resolution : 180 Resolution Unit : inches Exif Image Width : 400 Exif Image Height : 300 Focal Plane X Resolution : 8114.285714 Focal Plane Y Resolution : 8114.285714 Focal Plane Resolution Unit : inches Image Width : 640 Image Height : 480 If needed, the original image can be obtained from: here=http://www.pythonchallenge.com/pc/return/evil1.jpg wget --user=huge --password=file $here

    Read the article

  • MS Excel: Can I link images using a relative path?

    - by Port Islander 2009
    I am working on an MS Excel document that contains a lot of (around 200) images. They are currently saved within the document, so the file becomes huge and working gets very slow. Linking the pictures without saving them works very well - I now have the Excel document and a folder "pictures" next to it that contains all my image files. However, when I move the document and the folder to a new location, all my pictures disappear. This seems to be because Excel saves the link information as absolute paths. (Update: Actually, according to this thread, Excel stores the link information as relative paths as well. Now I really don't know why my links break down..) Is there a convenient way to save them as relative paths or have Excel automatically update the path information? Update: It's important that the images get displayed on the sheet and can be printed. I am working with Microsoft Excel for Mac 2008 and 2011. I really appreciate your help.

    Read the article

  • Graphing per-user CPU usage on a Linux machine

    - by mart1n
    I want to graph (graphical output would be great, i.e. a .png file) the following situation: I have users A, B, and C. I limit their resources so that when all users run a CPU intensive task at the same time, those processes will use 25%, 25%, and 50% of CPU. I know I can get the real-time stats using top but have no idea what to do with them. I've searched through the huge top man page but haven't found much on the subject of outputting data that can be graphed. Ideally, the graph would show a span of maybe 30 seconds. Any ideas how to achieve this?

    Read the article

  • Lots of artifacts while streaming HD content with VLC 0.9.9 on CentOS

    - by Zsub
    I'm trying to stream (multicast) a x264 encoded file using VLC. This in itself succeeds, but the stream has a huge lot of artifacts. This seems to suggest that the data cannot be transported fast enough. If I check network usage, though, it's only using about 15 mbit. I have a similar SD stream which functions perfectly. I think I could improve stream performance by not streaming the raw data, but I cannot seem to get this working. It seems that on keyframes all artifacts are removed for a short while (less than a second). This is the command I use: vlc -vv hdtest.mkv --sout '#duplicate{dst=rtp{dst=ff02::1%eth1,mux=ts,port=5678,sap,group="Testgroup",name="TeststreamHD"}}' --loop Which is all one long line.

    Read the article

  • writting becomes slow after few writes

    - by user1566277
    I am running an embedded Linux on arm with a SD-Card. While writing huge amounts of data I see bizarre effects. E.g, when I dd a 15 MB file few times, it writes the file (normally) in less than 2 Secs. But After lets say 3-4 times it takes sometimes 15 to 30 Seconds to write the same file. If I sync after writing the file, then this does not happen but sync takes long time too. If there is enough gap between writing two files than presumably kernel syncs itself. How can I optimize the whole performance so that write should always finish inside 2 Seconds. The File system I am using is ext3. Any pointers?

    Read the article

< Previous Page | 52 53 54 55 56 57 58 59 60 61 62 63  | Next Page >