Search Results

Search found 30347 results on 1214 pages for 'jamie read'.

Page 667/1214 | < Previous Page | 663 664 665 666 667 668 669 670 671 672 673 674  | Next Page >

  • Restrict subversion to only allow certain functions

    - by Farseeker
    I'm in a bit of a situation. We have our private subversion server that we use for development, but one of our government clients is requesting access to our commit logs so that they can get an up-to-date picture of what we've been doing on the system. I don't have a problem with them reading our commit logs, but what I do have a problem with is them having access to our source code - they can't have read or write. The obvious solution is to do an svn log ourselves and give them an export, but they want direct SVN access as they apparently have an auditing solution that will import the svn log command automagically. So, is there a way I can set up access to a subversion repo and deny them access to everything except svn log? I don't care if I have to set up a virtualhost just for this, but it has to be done over http(s). We're also using LDAP for authentication if that makes any difference.

    Read the article

  • Reducing the size of the EDB file.

    - by Toby
    I have hit an issue on a MS SBS machine where every morning the datastore for the exchange mailboxes dismounts itself. We believe the issue is that it has grown too large over time and needs cut down a bit. As part of this we have removed (purged) some mail files that were no longer needed, which should have given us a saving of roughly 3GB (more than enough saving for what we need). So I deleted the mailboxes, then purged them and noticed that the .edb file was still reporting the same size, I dismounted and remounted it to see if that would have any effect but it did not. Am I missing a step? I have read online that you can run offline defrag on the file but that seems to only save you a small amount of whitespace. Any help would be greatly appreciated.

    Read the article

  • process ksoftirqd consumes permanent 15% CPU load [closed]

    - by markus
    Possible Duplicate: Anyone else experiencing high rates of Linux server crashes during a leap second day? The process ksoftirqd/0 uses permanent 15% CPU on our debian squeeze server. 4 root 20 0 0 0 0 R 15.0 0.0 850:59.17 ksoftirqd/0 I already read that this can have various reason like Full harddisk or high network traffic. In our case we do have more or less low network traffic and enough space on hard disk. How can I analyse what causes ksoftirqd/0 to use permanently 15% CPU?

    Read the article

  • Safe to use high port numbers? (re: obscuring web services)

    - by sofakng
    I have a small home network and I'm trying to balance the need for security versus convenience. The safest way to secure internal web servers is to only connect using VPNs but this seems overkill to protect a DVRs remote web interface (for example). As a compromise, would it be better to use very large ports numbers? (eg. five digits up to 65531) I've read that port scanners typically only scan the first 10,000 ports so using very high port numbers is a bit more secure. Is this true? Are there better ways to protect web servers? (ie. web guis for applications)

    Read the article

  • Windows/Samba connection error

    - by Gomibushi
    I have a Linux fileserver serving up /home for linux and windows users. I was able to connect from my windows client, but not from a DC. Then suddenly I could connect from the DC too. The linux servers run Centrify clients, and as such are part of the domain. All on same subnet. This is what the the log.smbd says, repeatedly: [2010/02/11 11:25:57, 0] lib/util_sock.c:read_data(534) read_data: read failure for 4 bytes to client 192.168.200.3. Error = Connection reset by peer On Windows it appeared as an "unknown error". EDIT: the error code is "0x80004005". We are developing a system depended on the samba share, and are worried this will appear again. It would be nice to pin point the root of this. Any ideas what this might be? Places to look?

    Read the article

  • Imap subfolders not shown in list command - perhaps acls misconfiguration

    - by mschenk74
    My goal is to copy the whole folder structure with all mails from one imap account to another. The tool I am using for this is imapcopy (the java based version from code.google.com since the unix/linux tool packaged with debian doesn't support imaps). Now, there is one problem: The tool only copies the top-level folders and not the nested ones. To narrow down the problem I have downloaded the source code of imapcopy and debugged into the code. There I noticed that the folder.list() (which is mapped to the list "%" imap command) returns an empty list. But when I do a getFolder(<subfoldername>) I can access those subfolders. After reading some documentation about the features of imap I think that the problem might be some misconfigured ACLs which prohibit the listing of those folders but allow to read ad write to them. How should I check this ACLs? which tools do I need for this task?

    Read the article

  • How does the performance of pure Nginx compare to cpNginx?

    - by jb510
    There is now a Cpanel plugin to fairly easily setup Nginx as a reverse proxy on a Cpanel/Apache server. I've been simultaneously interested in setting up my first unmanaged VPS and my first Nginx server and as a masochist figured why not combine the two. I'm wondering however if it's worth setting up a pure Nginx server vs trying out cpNginx on Apache? My goal is solely to host WordPress sites and while what I've read raves about Nginx's is exceptional ability serving static at least as a reverse proxy, I am unclear if there is substantial benefit to running a pure nginx with eAccelorator over cpNginx on Apache for dynamic sites? Regardless I'll be running W3TC on all sites to cache content, but am still interested if there are big CPU reductions running PHP scripts under pure Nginx over cpNginx?

    Read the article

  • Moving a File While It's In Use -- How Does it work?

    - by Zaz
    I've noticed that on non-windows OS.... ie linux/mac i can do things like: - Send a zip to a friend over aim - Delete the file while it's in transfer And the transfer does not fail. Or, I can do operations like.. - start a movie - erase the file - the movie still plays to completion (read from disk, not just buffered in memory) Although the files are being "deleted", as i mentioned, they are actually being moved to a different location on the file system... ie a Trash directory or something. So it seems to me like the OS uses a pointer @ the file that is updated when it moves rather than accessing the files directly. Can anyone shed some light on how this AWESOME capability is actually implemented? I'm not even sure what to google to learn more about it. thank you.

    Read the article

  • Is is possible to slipstream DotNet 4 into Windows XP ?

    - by user43300
    I tried to use nLite but this program is no longer updated. And it seems it has no capability of dealing with dotNet. On my instalation CD there is DOTNETFX folder with version 1.1, but it seems even this old version is not installed with system. I also read somewhere that it is possible to download dotNet as system update, that is a file with name that begins wih KB.... , so that file could be properly digested by nLite ? Any help in this matter ?

    Read the article

  • Lenovo N100 and Samsung SSD 830, shouldn't this go faster than 100MB/s?

    - by Jook
    I recently upgrated my Lenovo N100 0768 with a Samsung SSD 830 - specified to support 520 MB/s read and 320MB/s write. However, having only SATA1 with max. of 150MB/s speed, it has to run in a slower mode - of course. But, shouldn't it be more on the edge of this limit than ranging between 90-115MB/s and averaging around 103MB/s? Or is this really as much as I can expect? Tested with HD-Tach. Has anyone comparable results on a SATA1 controller with a SSD drive faster than SATA1? Preferably similar to the Samsung 830?

    Read the article

  • compiling the linux kernel

    - by user482819
    Just for learning, I have recompiled the linux kernel with different options, installed and boot from it. It was both instructive and straight forward. However, I was overwhelmed by the big number of options available. My questions are: 1.- Does it make sense to spend time trying to optimize the linux kernel for my particular laptop? Will it make a significant improvement? 2.- Is there any tool that can read the configuration of my computer and suggest a config? Thanks, H

    Read the article

  • Windows 7 ISO not booting from DVD drive?

    - by chum of chance
    I've downloaded Windows 7 x64 Pro Upgrade from the volume license center. I've also downloaded the Windows 7 USB/DVD download tool. I'll burn it using this tool and while it will show up in Explorer as if the burn worked correctly, when I try to boot to the DVD on startup, I get a "No operating system found" error (this is a newer Lenovo). My Ubuntu Live CD works fine, am I missing something? I read that there's problems if you try to burn a 64bit copy from a 32bit computer, but I'm currently running Win7x64, so that shouldn't be a problem. I've also tried burning with ImgBurn and Windows 7 built ISO burner, no success. Obviously there's a step I'm missing here, what gives?

    Read the article

  • let CRON send emails through SMTP ( debian squeeze )

    - by supernova
    i would like to send emails whenever a cronjob has completed, i read that this is possible with exim4. in /etc/alias i added the line myuser: [email protected] in /etc/exim4/update-exim4.conf.conf i set dc_smarthost='smtp.myserver.ip::25 and in /etc/exim4/passwd.client i set smtp.myserver.ip:[email protected]:mypassword my problem is that i can't see any login at my mailserver, and in the exim logs i saw a few lines with 2012-10-13 09:17:01 1TMvy1-0001fp-F2 ** [email protected] R=nonlocal: Mailing to remote domains not supported 2012-10-13 09:17:01 1TMvy1-0001fr-JE <= <> R=1TMvy1-0001fp-F2 U=Debian-exim P=local S=17426 are there any additional config settings i have to set? edit : i solved the prev. error by running dpkg-reconfigure, but now i'm facing the following error <root@debian> R=dnslookup T=remote_smtp defer (-53): retry time not reached for any host

    Read the article

  • Safe to use high port numbers? (re: obscuring web services)

    - by sofakng
    I have a small home network and I'm trying to balance the need for security versus convenience. The safest way to secure internal web servers is to only connect using VPNs but this seems overkill to protect a DVRs remote web interface (for example). As a compromise, would it be better to use very large ports numbers? (eg. five digits up to 65531) I've read that port scanners typically only scan the first 10,000 ports so using very high port numbers is a bit more secure. Is this true? Are there better ways to protect web servers? (ie. web guis for applications)

    Read the article

  • Setup for NEW HDD?

    - by Ed Markey
    Building my FIRST computer. Will start with at least - one HDD, I purchased a WD SATA HDD a 10,000RPM VelociRaptor 500GB. I have read that if you use one of the settings, can't remember what it's called, that it would make an HDD rund at IDE speeds??? Could I or should I run it in RAID instead? Is RAID only for more than one HDD/ssd? I would like to get like a 64GB SSD later to maybe speed things up working together. Thanks

    Read the article

  • Received a RAMPAGE IV FORMULA MOBO. Possible CPU Socket problems

    - by Tantan
    I recently received a Rampage IV Formula MOBO in a trade. From everything I read online this seems like a pretty nice piece of hardware. There's one thing I'm worried about though. It appears as though the socket where the CPU goes might have some bent/missing pins. I'm not exactly a computer expert and I wanted to know if this could be any sort of problem. I would post a picture but I apparently don't have enough rep for that... Also, if I did want to move my current rig onto the new MOBO would it even be worth it? MY CURRENT COMPUTER SPECS i3-3220 CPU 8GB DDR3 RAM GeForce GTX460 GPU With a Biostar MOBO!

    Read the article

  • The authenticity of host “host” can't be established

    - by Candroid
    I'm running a web app on a Linux server which connects to other servers. When I run the project on my Play framework on loclhost it runs fluently. When I run it on my Linux server I get the above message 3 times, one for each server. I read a post about it where it says that it is a man in the middle warning and if I write yes it should work. But though a write yes, nothing happens and the app doesn't run, and the error message keeps popping up. I tried creating private and public keys and add them to the authorized_keys file, but it didn't work either. What should I so to run my app?

    Read the article

  • Why might one hard disk perform slower than another?

    - by Styne666
    I have just bought two WD 3TB Reds (WD30EFRX) for a FreeNAS box and whilst doing burn-in testing it seems like one is consistently taking about 10% longer than the other. So far I've done: a dd read test of the whole device, a long SMART test and it's currently halfway through a badblocks -wvs. The second device is lagging behind the first on all of them. I'm running these commands on Debian stable in two Konsole tabs. Is there a reason this could be considered normal behaviour or is it worth running the tests independantly? They're both plugged in to the LSI 2308 (IT mode) on a Supermicro X10SL7-F.

    Read the article

  • What's a worthwhile test for a new HD?

    - by Michael Kohne
    I work for a company that uses standard 2.5" SATA HD's in our product. We presently test them by running the Linux 'badblocks -w' command on them when we get them - but they are 160 gig drives, so that takes like 5 hours (we boot parted magic onto a PC to do the scan). We don't actually build that many systems at a time, so this doable, but seriously annoying. Is there any research or anecdotal evidence on what a good incoming test for a hard drive should be? I'm thinking that we should just wipe them with all zeros, write out our image, and do a full drive read back. That would end up being only about 1 hour 45 minutes total. Given that drives do block remapping on their own, would what I've proposed show up any infant mortality just as well as running badblocks?

    Read the article

  • Can files deleted on an ecnrypted drive be restored?

    - by roddik
    Hi. There are ways to restore files, deleted from the system by default, I'm not sure about the way they work but I guess thet read content, that has not been overwritten. On the other hand, there are programs (e.g. TrueCrypt), that encrypt disks, claiming that it wouldn't be possible to tell apart random data and file contents on such a disk without a password. Therefore I think that files, deleted from such disks can't be restored. Is that correct? I know one way to find out would be to try it, but there is a possibility, that I would just pick the wrong restoring software. Moreover, I'm more interested in theorethical explanation why it would/wouldn't be possible. Thanks

    Read the article

  • Why does Chrome show overlapping text?

    - by dog44wgm
    In Chrome, news articles at: http://www.theprovince.com with a leading photo and caption show the caption text overlapped with the body text. I have an image but as a new user here I'm not allowed to upload it. It happens at that site almost always, here's an example from today: http://www.theprovince.com/sports/Canucks+Blackhawks+collision+Titanic+proportions/5721421/story.html It rarely happens elsewhere. The same link works fine in Internet Explorer so I'm guessing it's a Chrome issue. It's been like this for many months, I read the site almost everyday. I click on "Print this Article" to get a proper look at it, but it's annoying, hope someone has the answer. Thanks in advance.

    Read the article

  • WGet a Page that Requires Logging in

    - by Synetech inc.
    I’m trying to figure out a way to use WGET or a similar tool so that I can schedule a web page to be downloaded regularly as a sort of updating log. The problem is that the page requires that I be logged in otherwise I get a different page, generic. Further, the page does not take login information as GET parameters in the URL, it uses POST to log in on the login page and cookies to save the login information that’s read by the regular page. I’m currently using GNU Wget 1.10.2 for Windows. I’ve tried using WGET’s cookie functionality but have had mixed results, usually skewing towards it not working. Can anyone please advise on a way to accomplish this? Thanks a lot.

    Read the article

  • What's an alternative to using public folders (in Outlook)?

    - by Ivo Flipse
    My colleagues abuse our mail servers public folders to store (old) emails so that everyone can read them using IMAP. I'm looking into good alternatives after reading this Tech Republic article: "10 reasons why you should begin phasing out Exchange public folders" The most important thing they need is access to emails from multiple computers without overloading our network. So do you have any suggestions for alternatives? If there's a nice combination with some CRM system it would be interesting too. Note: this doesn't have to be freeware, usability and efficiency are more important. The solution has to be Windows 32 bit only

    Read the article

  • Changing Chrome Working Directories on Mac clients

    - by Jason
    We're using Mac OSX 10.6.8 with NFS for home directories. Chrome seems to not like this, for a user who is logged in more than one machine. AFP was ok, but for other reasons we need NFS. It only allows you to start one instance. Ie, userA logged in to Mac1, Mac2. Can only have chrome open on Mac1 OR mac2, not both. I have read at this thread, changing working directory, and sym link can fix... but I can't find the details. Can someone help?

    Read the article

  • Copying large file from SD to HDD via USB failing on Ubuntu

    - by Kent Boogaart
    Hi, I'm attempting to copy some large files from my camera (Canon EOS 500D) to my laptop, which is running 64 bit Ubuntu 9.04. I am using USB to connect the two devices. For most files, it is simply a matter of control-C and control-V. I have done this successfully many times with both photos and small movies (eg. 180MB). However, when I attempt to do this with very large files (eg. 3GB), the copy seems to start with a lot of activity both on the camera and laptop, but after 10 minutes or so the camera is automatically unmounted and the copy fails to complete. I have read that this might be due to the device not mounting as a mass storage device, but I cannot see any obvious way for me to change this behavior. Can anyone offer any direction here? I'll get a USB card reader if necessary, but I'd prefer to be able to just plug my camera in. Thanks

    Read the article

< Previous Page | 663 664 665 666 667 668 669 670 671 672 673 674  | Next Page >