Monthly Archives

Articles indexed in September 2012

Page 45/470 | < Previous Page | 41 42 43 44 45 46 47 48 49 50 51 52  | Next Page >

  • Subversion all or nothing access to repo tree

    - by Glader
    I'm having some problems setting up access to my Subversion repositories on a Linux server. The problem is that I can only seem to get an all-or-nothing structure going. Either everyone gets read access to everything or noone gets read or write access to anything. The setup: SVN repos are located in /www/svn/repoA,repoB,repoC... Repositories are served by Apache, with Locations defined in etc/httpd/conf.d/subversion.conf as: <Location /svn/repoA> DAV svn SVNPath /var/www/svn/repoA AuthType Basic AuthName "svn repo" AuthUserFile /var/www/svn/svn-auth.conf AuthzSVNAccessFile /var/www/svn/svn-access.conf Require valid-user </Location> <Location /svn/repoB> DAV svn SVNPath /var/www/svn/repoB AuthType Basic AuthName "svn repo" AuthUserFile /var/www/svn/svn-auth.conf AuthzSVNAccessFile /var/www/svn/svn-access.conf Require valid-user </Location> ... svn-access.conf is set up as: [/] * = [/repoA] * = userA = rw [/repoB] * = userB = rw But checking out URL/svn/repoA as userA results in Access Forbidded. Changing it to [/] * = userA = r [/repoA] * = userA = rw [/repoB] * = userB = rw gives userA read access to ALL repositories (including repoB) but only read access to repoA! so in order for userA to get read-write access to repoB i need to add [/] userA = rw which is mental. I also tried changing Require valid-user to Require user userA for repoA in subversion.conf, but that only gave me read access to it. I need a way to default deny everyone access to every repository, giving read/write access only when explicitly defined. Can anyone tell me what I'm doing wrong here? I have spent a couple of hours testing and googling but come up empty, so now I'm doing the post of shame.

    Read the article

  • Log centralization, display, transport and aggregation at scale v2

    - by Eric DANNIELOU
    This is a duplicate question of Log transport and aggregation at scale and http://stackoverflow.com/questions/1737693/whats-the-best-practice-for-centralised-logging, but the answers might differ now : The softwares described in 2009 may have changed since (for example Octopussy evolved from version 0.9 to 1.0.5). Rsyslog has become the default on most linux distro. Requirements have changed (security, software configuration management, ...). I'd like to ask the following questions : How do you centralize, display and archive system logs? How would you like to do it now if you had to? Most linux distro use rsyslog nowadays, which can provide reliable log transport. But some older unices, network devices and maybe windows box still use old udp rfc-style transport. How did you manage to get reliable transport? Storing logs for a few months can represent a huge amount of disk space. How do you store them? rdbms? Compressed and encrypted text files?

    Read the article

  • How to find out if my IIS is x64 or x86?

    - by PaparazzoKid
    As the title reads, I'm about to download and install URL Rewrite Go Live extension on my Windows Server 2008 R2 dedicated server, and I need to find out which version of URL Rewrite I should be downloading, x64 or x86. The information that came with the server when I started to rent it said it was 64-bit but when I look at my C: directory I have two Program Files directories, one titled "Program Files" and the other "Program Files (86)" - so this has confused me somewhat. Can anybody suggest how to find this out?

    Read the article

  • Low-cost, Flexible Log Aggregation [closed]

    - by Dan McClain
    I'm starting to have quite the collection of Ubuntu VMs that I must manage. I'm starting to investigate Puppet for managing the configuration of all of them, and apticron to let me know what's out of date. But the issue I feel I should deal with sooner than later is log aggregation. I'd like to stay in the free/open source realm for now, seeing that we don't have much budget for something like splunk yet. In addition to syslog, I would like to collect application specific logs (We are running different apps on different machines, from nginx+passenger for rails, to Apache+Tomcat for java, to PHP for expression engine, and mysql/postgresql database server), so that we can analyze the relavent data. For now, I'm just looking to get all the logs one place.

    Read the article

  • Splunk is fantastically expensive: What are the alternatives? [closed]

    - by samsmith
    Possible Duplicate: Alternatives to Splunk? This has been discussed, but it has been several months, so it may be time to revisit it: Earlier discussion RE Splunk alternatives For the record, Splunk rocks. But the pricing is simply beyond what we can consider (When I spoke with Splunk today, the cost for a system to index 5gb/day of data is over $30,000.) That is more than we spend on SQL Server (by a large multiple), more than we spend on a rack of servers (by a multiple), etc. etc. The splunk sales team is correct (that for $30K we get more value and functionality than if we spend the same building our own system), but it doesn't matter. The splunk cost is simply too high (by a multiple). Soooooo, we are looking around! Is anyone out there building a splunk like system? Our basic need: Able to listen for syslog messages on multiple udp ports Able to index the incoming data in an async way Some kind of search engine Some kind of UI An API to the search engine (to embed in our console) We currently need to index 3-5gb/day, but need to be able to scale to 10gb/day or more. We do not need a lot of history (30 days is fine). We use Windows 2008 and 2003 servers. Thanks for your thoughts! UPDATE: We spent two weeks researching commercial and open source options. Our conclusion: Write our own (we are a software company... we know how to write things). We built a great system built on mongodb and .NET that gives us the functions we needed from MongoDB in about one engineering week. We have now completed our implementation. We use two Mongodb servers (master and slave), and are able to log and index any amount of log data (5gb/day, 15gb/day, etc), limited only by disk space. OBSERVATIONS: This space needs a solid solution that is $1000-3000 flat rate. The licensing models used by the commercial firms are based on a "milk the data center ops guys" models. That is their right (of course!), but it leaves a HUGE space open for someone to come in underneath them. My guess is that in another year or two there will be a good open source solution that will be really usable. Thank you all for your input (even if it was self promotion).

    Read the article

  • NTBackup Error: C: is not a valid drive

    - by Chris
    I'm trying to use NtBackup to back up the C: Drive on a Microsoft Windows Small Business Server 2003 machine and get the following error in the log file: Backup Status Operation: Backup Active backup destination: 4mm DDS Media name: "Media created 04/02/2011 at 21:56" Error: The device reported an error on a request to read data from media. Error reported: Invalid command. There may be a hardware or media problem. Please check the system event log for relevant failures. Error: C: is not a valid drive, or you do not have access. The operation did not successfully complete. I'm using a brand new SATA Quantum Dat-72 drive with a brand new tape (tried a couple of tapes). I carry out the following: Open NTBackup Select Backup Tab Tick the box next to C: Ensure Destination is 4mm DDS Media is set to New Press Start Backup Choose Replace the data on the media and press Start Backup NTBackup tries to mount the media Error Message shows: The device reported an error on a request to read data from media. Error reported: INvalid command. There may be a hardware or media problem. Please check the system event log for relevant failures. On checking the log I find the following: Event Type: Information Event Source: NTBackup Event Category: None Event ID: 8018 Date: 04/02/2011 Time: 22:02:02 User: N/A Computer: SERVER Description: Begin Operation For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp. and then; Event Type: Information Event Source: NTBackup Event Category: None Event ID: 8019 Date: 04/02/2011 Time: 22:02:59 User: N/A Computer: SERVER Description: End Operation: The operation was successfully completed. Consult the backup report for more details. For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp.

    Read the article

  • How can I erase the traces of Folder Redirection from the Default Domain Policy

    - by bruor
    I've taken over from an IT outsourcer and have found a struggle now that we're starting a migration to windows 7. Someone decided that they would setup Folder redirection in the Default Domain Policy. I've since configured redirection in another policy at an OU level. No matter what I do, the windows 7 systems pick up the Default Domain Policy folder redirection settings only. I keep getting entries in the event log showing that the previously redirected folders "need to be redirected" with a status of 0x80000004. From what I can tell this just means that it's redirecting them locally. Is there a way I can wipe that section of the GPO clean so it's no longer there? I'm hesitant to try to reset the default domain policy to complete defaults. ***UPDATE 6-26 I found that the following condition occurred and was causing the grief here. I've already implemented the new policies for clients, and for some reason, XP was working great, 7 was refusing to process. The DDP was enforced. Because of this, and the fact that the folder redirection policies were set to redirect back to the local profile upon removal, it was forcing clients to pick up it's "redirect to local" settings. Requirements for to recreate the issue. -Create a new test OU and policy. -Create some folder redirection settings, set them to redirect to local upon removal -Remove settings on that GPO -Refresh your view of the GPO and check the settings. -You'll notice that the settings show "not configured" entries for folder redirection. -Enforce this GPO -Create another sub-OU -Create a GPO linked to this sub-ou and configure some folder redirection settings. -Watch as the enforced GPOs "not configured" setting overrides the policy you just defined. I've had to relink the DDP to all OU's that have "block inheritance" enabled, and disable the "enforced" option on the DDP as a workaround. I'd love to re-enable enforcement of the DDP, but until I can erase the traces of folder redirection settings from the DDP, I think I'm stuck.

    Read the article

  • Recommended programming language for linux server management and web ui integration

    - by Brendan Martens
    I am interested in making an in house web ui to ease some of the management tasks I face with administrating many servers; think Canonical's Landscape. This means doing things like, applying package updates simultaneously across servers, perhaps installing a custom .deb (I use ubuntu/debian.) Reviewing server logs, executing custom scripts, viewing status information for all my servers. I hope to be able to reuse existing command line tools instead of rewriting the exact same operations in a different language myself. I really want to develop something that allows me to continue managing on the ssh level but offers the power of a web interface for easily applying the same infrastructure wide changes. They should not be mutually exclusive. What are some recommended programming languages to use for doing this kind of development and tying it into a web ui? Why do you recommend the language(s) you do? I am not an experienced programmer, but view this as an opportunity to scratch some of my own itches as well as become a better programmer. I do not care specifically if one language is harder than another, but am more interested in picking the best tools for the job from the beginning. Feel free to recommend any existing projects that already integrate management of many systems into a single cohesive web ui, except Landscape (not free,) Ebox (ebox control center not free) and webmin (I don't like it, feels clunky and does not integrate well with the "debian way" of maintaining a server, imo. Also, only manages one system.) Thanks for any ideas! Update: I am not looking to reinvent the wheel of systems management, I just want to "glue" many preexisting and excellent tools together where possible and appropriate; this is why I wonder about what languages can interact well with pre-existing command line tools, while making them manageable with a web ui.

    Read the article

  • Batch Scripting - Listing files with a specific amount of characters in file name

    - by Jane
    I'm creating a batch script for a class and I've hit a roadblock I have to list all text files whose names are up to seven characters long on the whole c: drive - make the listing output in a wide formant - then append to Batch script file output.txt So far I have -- dir c:*txt/w/o/s/p c:/"My Batch Script File Assigment"/"Output"/"Batch Script File Output Data".txt The above does everything except limit the search to files with only 1-7 characters in their name. If anyone could point me in the right direction I would really appreciate it!

    Read the article

  • Make ZoneAlarm stop pausing my C programs when I run them

    - by rMaero
    I'm using Dev-C++ to develop some console apps to study. When my program tries to run system("PAUSE"); ZA stops it and asks me to allow or deny it. I check the "always" box but it seems that every time I compile it, it generates a new exe file so every time I run it ZA pops up. Of course the simplest solution is to disable it or deal with it :-P but I'm not eager for any of both. Any suggestions? thanks in advance!

    Read the article

  • 1080p playback on external monitor

    - by xibalban
    My system (netbook) specs: 1.6 GHz dual core atom N2600 2 GB DDR3 RAM Integrated GMA3600 GPU Win 7 Professional 32 Bit Daum Pot Player V-1.5 The issue: I downloaded a 1.33 GB sized 1920 x 1080 HD documentary (file extension: mp4) and it plays without a hitch using pot player. However, when I connected an external display (an LED Full HD 24" monitor) using D-Sub, the video plays fine but the player controls become very unresponsive. For instance, it takes about 40 seconds to exit full screen while playing. I tried: Installed/updated all the latest monitor drivers Updated the VGA drivers for GMA3600 No other programs ran while the video played, with most unwanted services and background processes disabled/killed. CPU usage goes only up to 49% when the video plays Question: What may I do in order to improve the player response/video playback experience? Do I need to play with (or change) Pot Player's default video decoders, etc?

    Read the article

  • jump to page of a pdf in google docs / drive / apps

    - by Aaron - Solution Evangelist
    i want to jump to a specific page of a pdf file via the google docs via the editor url https://docs.google.com/file/d/xxx/edit or the embed url https://docs.google.com/file/d/xxx/preview i am not looking to use the http://docs.google.com/gview?url= referenced in the stackoverflow question how to open specific page on Google's docs viewer as i want to do this for documents where authentication is required the the document is not available via public url. is there some way of appending an anchor (i would have expected it to be https://docs.google.com/file/d/xxx/preview#10) or a query (e.g. https://docs.google.com/file/d/xxx/preview?page=10) to the google docs / drive / apps viewer?

    Read the article

  • Is it possible to mount a disk image, created with dd, to a directory on a mounted external usb hdd?

    - by Keeper Hood
    I have an image of my home (/dev/sda3) partition, which I've created using the "dd" command. dd if=/dev/sda3 of=/path/to/disk.img I've deleted the home partition via gparted in order to enlarge my /dev/root partition. Then I've recreated the /dev/sda3 partition which is smaller in size then the one I've backed up to the image. I was wondering since I have a 2TB external HDD, could it be possible to mount my backed up image on the external HDD and then copy the files into the /home directory. Since the external HDD would be already in a "mounted state", I'm unsure whether this is a good idea, mounting on a mounted device. I'm running Slackware 13.37 (64bit). used ext4 on all the partitions. resized the root partition with gparted live cd. I've tried mount -t ext4 /path/to/disk.img /mng/image -o loop It gave me an fs error (wrong fs type, bad option, bad superblock on dev/loop/0) Then i did dmesg | tail which outputs: EXT4-fs (loop0) : bad geometry: block count 29009610 exceeds size of defice (1679229 blocks) I have no idea what to do, I want to restore my /home data from the image I've backed up.

    Read the article

  • VirtualBox - split partitioned VDI into separate VDIs

    - by mathematical.coffee
    I'm very new to VirtualBox. I set up an Arch Linux VM and a Ubuntu VM (Ubuntu host), both sharing the same .vdi like so (I had in my mind a dual-boot situation): VDI file (25GB) |- /dev/sda1: 5GB (Arch Linux) |- /dev/sda2: [Ubuntu] |- /dev/sda5 (swap, 1GB) |- /dev/sda6 Ubuntu /, 9GB |- /dev/sda7 Ubuntu /home, 10GB I've now realised that I don't want a dual-boot-type setup, I'd rather boot each machine independently (my initial thought was to share /home between Ubunto and Arch). So, my question: Can I split /dev/sda1 and /dev/sda2 each to their own .vdi files so I can use them as completely separate machines? I'd rather not have to re-install either Arch (because it took me ages to work it out!) or Ubuntu (because I've already done a few GB of updates and don't want to redo them). I haven't been able to find anything about this - most questions I see are about converting a .vdi to a partition on the host, or splitting a .vdi into multiple smaller files (that are not independent), or converting a partition on the host to a .vdi file. cheers.

    Read the article

  • Unable to use NTFS partition for Dropbox in Linux

    - by Cristian
    Dropbox won't let me choose the sync folder inside a NTFS partition. First thought I had was mounting and its permissions (the Dropbox installer does let me choose my linux home as the Dropbox home). After searching and trying several other lines, the partition is mounted via fstab with these settings: /dev/sda5 /mnt/documents ntfs-3g uid=1000,gid=100,dmask=027,fmask=137 0 0 I can read and write in the partition, here is a ls output: 24 drwxr-x--- 1 tuxcayc users 24576 Sep 2 06:42 documents I'm using an Arch-based distro (Manjaro) and Dropbox installed via yaourt. I guess it's still some issue with mounting permissions. Any help is appreciated, Thanks.

    Read the article

  • Run script on login with ssh

    - by user912447
    I have a feeling this is quite easy to do but every solution found on google has to do with adding a script to be run whenever someone logs into the machine. What I am looking for is a way to run a script when only I log into the machine. I ssh into a shared computer and need to have it load a couple modules for me and I imagined the easiest way to do this would to just run a script on login. Is there a simple way?

    Read the article

  • Wi-fi signal with keeping the internet cable

    - by daGrevis
    So the situation is that I have Ethernet cable which provides internet to my computer. Thing I want is to have wi-fi connection in my house and Ethernet cable (like I have before) to use for my PC. I will use wi-fi for my laptop and mobile phone. I think I need router for that and I'm looking at Asus RT-N16 (suggested in Coding Horror) for it, but I am not sure. Is it the right thing for me and I will be able to get wi-fi signal and keep the Ethernet cable? I guess the system will be that current cable goes into router, router provides wi-fi signal and gives back new cable... or something like that. Thanks in any advice! And sorry if topic isn't in right site.

    Read the article

  • HP C4400 is paper out

    - by borges
    I have a HP C4400 printer and it claims to be out of paper (which is not true) and don't even try to pull the paper. I have already followed these instructions with no success: http://h10025.www1.hp.com/ewfrf/wc/document?docname=c00786157&cc=us&lc=en&dlc=en&product=3418705 Two strange things: I used the printer successfully minutes before this problem. It prints its test page successfully (by pressing the cancel and color buttons simultaneously). How can I fix this problem? PS: Sorry about the creepy english.

    Read the article

  • PDF printer which correctly embeds EPS into PDF

    - by Alexey Popkov
    I need to convert to PDF a Word document containing embedded vector EPS images (by printing to PDF printer - I use Word 2003). Several years ago I tested some of commercial and free PDF printers and found none, with except to Acrobat Distiller, which embeds in the generated PDF file real PostScript content of the EPS image instead of the preview showed by Word. Has the situation changed from that time? Do you know any free or commercial PDF printer which handles embedded EPS correctly? UPDATE Good thread about EPS handling in different versions of Word: http://forums.adobe.com/thread/439881

    Read the article

  • Automatic sort for excel worksheet

    - by Joseph
    I want to create a to-do list in Excel that automatically sorts the to-do entries in a list, in order of ones to do first (closest deadlines). I would also like a section that shows the tasks for today and another for high-priority tasks coming up within a week. I have not programmed in Excel before. I know Python and JavaScript, but want an Excel solution that runs inside Excel (maybe using VBA, the Excel programming language). Is this sort of thing possible in Excel?

    Read the article

  • Can I delete the OEM partition on the new Dell XPS 15?

    - by timepilot
    My new Dell XPS 15 L521X just arrived. I need to set this up to dual boot Linux. Sadly, the system comes with four primary partitions. I can't do a clean install at the moment, so one of the partitions will have to be deleted before I can install Linux. The layout is as follows: OEM: 39mb Hibernation: 8gb OS: 457gb Recovery: 12gb Obviously, I can't delete the Hibernation and OS partitions (will shrink to make space for Linux) and I'd like to keep the recovery partition if possible. So my question is what is on the small OEM partition? What functionality will I lose if I delete it?

    Read the article

  • How to recover data from Dell dimension 9150 RAID 0 on another system?

    - by Adam
    I have a Dell Dimension 9150 which has failed. I'm trying to recover the data. It had two SATA 250GB drives in RAID 0 configuration. I'm trying to use a shuttle PC running Windows7 to recover the data from the drives which contained an XP boot volume. I just want the data, it doesn't have to boot. What program would I need to rebuild / interrogate the drives (one of them is failing the onboard hardware test)? What drivers would I need to install? Windows7 sees the drive as one partition but doesn't see volume information. Ubuntu can see the drive as one partition and also can tell what the drive is called, but can't access the data Any help appreciated!! :)

    Read the article

  • Why does MS-DOS tells me I need extended memory and thinks a file is read-only? [closed]

    - by Jake Inc.
    I am running a .COM file on a MS-DOS 6.22 boot USB drive When I run it on my laptop the program works fine but when I run it on my desktop I get error 40 not enough extended memory. When I go to the memory tab in GUI I try to switch none to auto, but I get the error "This file might be read-only". It's not read-only, when I put it on my desktop I change the settings and the new settings are in a .pif I can't run .pif in MS-DOs so I need to Change the .exe not create a .pif. Change the amount of extended memory all files have on my MS-DOS. On my laptop there is no memory tab, the only real difference is my laptop is x64. Thanks for helping but I think teh x64 bit has nothing to do with it I dont eve nthink iits in 64 bit mode because Im using a boot USB. What i need to is listed above, thanks for helping.

    Read the article

  • vncviewer connection refused (61)

    - by coure2011
    I have a root access to VPS (centos 6). I have installed VNCServer using this guide line http://tournasdimitrios1.wordpress.com/2011/02/02/how-to-setup-vnc-server-on-centos-5-x-fedora-11/ Everything goes perfectly and server is running via termina. Now I am trying to connect to that server via vncviewer (mac os). but its giving me error Connection refused (61) I am providing only the IP address of the VPS, maybe I also needed port address? How to configure port on vncserver? or its something else?

    Read the article

  • Linux - Network Sharing a local NTFS usb drive

    - by Jonathan Rioux
    I have an external hard drive formated in NTFS which I would like to be able to access by the network. I want to make a network share out of it. I also have a Debian machine running in my house and I then got an idea. I want to plug in my external hard drive (usb) into my Debian machine, and make a windows share with it, maybe with Samba, so I will be able to access it from my Windows 7 laptop and see it as a network share. Additionally, how can I restrict specific folders of that network share, and allow only specific folders to specific users? For instance, I would like to give my girlfriend access to a folder of her name so she can put her files and so she wont be able to see the stuff in my folder...

    Read the article

< Previous Page | 41 42 43 44 45 46 47 48 49 50 51 52  | Next Page >