Search Results

Search found 29191 results on 1168 pages for 'joel in go'.

Page 506/1168 | < Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >

  • using svnadmin in a php script

    - by fabjoa
    Howdie Scenario: Allow developers to submit new application packages to a market server. Developers run a bash script which contains a cURL call to market server (localhost/market/submit/$app-name). The submit script on the server creates a new folder in existing svn server with the name of the submitted app. Script on dev side waits for HTTP to issue a success message and then do a svn checkout in dev local machine. Problem: The submit script on the market server failed to create new svn directory through code: echo `svnadmin mkdir -m 'added new package $package' http://localhost/market/packages/$package`; this does not echo nothing and when I go on http://localhost/market/packages, the folder has not been added and the revision number has not been incremented. I've tried from a terminal in market server chown root:www-data /usr/bin/svnadmin but still no luck. Somebody has come acrosss similar problem? Any solutions? Thanks! Profile: Linux/Ubuntu, apache subversion

    Read the article

  • Linux FHS: /srv vs /var ... where do I put stuff?

    - by wag2639
    My web development experience has started with Fedora and RHEL but I'm transitioning to Ubuntu. In Fedora/RHEL, the default seems to be using the /var folder while Ubuntu uses /srv. Is there any reason to use one over the other and where does the line split? (It confused me so much that until very recently, I thought /srv was /svr for server/service) My main concern deals with two types of folders default www and ftp directories specific application folders like: samba shares (possibly grouped under a smb folder) web applications (should these go in www folder, or do can I do a symlink to its own directory like "_/www/wordpress" - "/srv/wordpress") I'm looking for best practice, industry standards, and qualitative reasons for which approach is best (or at least why its favored).

    Read the article

  • Cross-platform distributed fault-tolerant (disconnected operation/local cache) filesystem

    - by Adrian Frühwirth
    We are facing a design "challenge" where we are required to set up a storage solution with the following properties: What we need HA a scalable storage backend offline/disconnected operation on the client to account for network outages cross-platform access client-side access from certainly Windows (probably XP upwards), possibly Linux backend integrates with AD/LDAP (permission management (user/group management, ...)) should work reasonably well over slow WAN-links Another problem is that we don't really know all possible use cases here, if people need to be able to have concurrent access to shared files or if they will only be accessing their own files, so a possible solution needs to account for concurrent access and how conflict management would look in this case from a user's point of view. This two years old blog posts sums up the impression that I have been getting during the last couple of days of research, that there are lots of current übercool projects implementing (non-Windows) clustered petabyte-capable blob-storage solutions but that there is none that supports disconnected operation nicely and natively, but I am hoping that we have missed an obvious solution. What we have tried OpenAFS We figured that we want a distributed network filesystem with a local cache and tested OpenAFS (which, as the only currently "stable" DFS supporting disconnected operation, seemed the way to go) for a week but there are several problems with it: it's a real pain to set up there are no official RHEL/CentOS packages the package of the current stable version 1.6.5.1 from elrepo randomly kernel panics on fresh installs, this is an absolute no-go Windows support (including the required Kerberos packages) is mystical. The current client for the 1.6 branch does not run on Windows 8, the current client for the 1.7 does but it just randomly crashes. After that experience we didn't even bother testing on XP and Windows 7. Suffice to say, we couldn't get it working and the whole setup has been so unstable and complicated to setup that it's just not an option for production. Samba + Unison Since OpenAFS was a complete disaster and no other DFS seems to support disconnected operation we went for a simpler idea that would sync files against a Samba server using Unison. This has the following advantages: Samba integrates with ADs; it's a pain but can be done. Samba solves the problem of remotely accessing the storage from Windows but introduces another SPOF and does not address the actual storage problem. We could probably stick any clustered FS underneath Samba, but that means we need a HA Samba setup on top of that to maintain HA which probably adds a lot of additional complexity. I vaguely remember trying to implement redundancy with Samba before and I could not silently failover between servers. Even when online, you are working with local files which will result in more conflicts than would be necessary if a local cache were only touched when disconnected It's not automatic. We cannot expect users to manually sync their files using the (functional, but not-so-pretty) GTK GUI on a regular basis. I attempted to semi-automate the process using the Windows task scheduler, but you cannot really do it in a satisfactory way. On top of that, the way Unison works makes syncing against Samba a costly operation, so I am afraid that it just doesn't scale very well or even at all. Samba + "Offline Files" After that we became a little desparate and gave Windows "offline files" a chance. We figured that having something that is inbuilt into the OS would reduce administrative efforts, helps blaming someone else when it's not working properly and should just work since people have been using this for years. Right? Wrong. We really wanted it to work, but it just doesn't. 30 minutes of copying files around and unplugging network cables/disabling network interfaces left us with (silent! there is only a tiny notification in Windows explorer in the statusbar, which doesn't even open Sync Center if you click on it!) undeletable files on the server (!) and conflicts that should not even be conflicts. In the end, we had one successful sync of a tiny text file, everything else just exploded horribly. Beyond that, there are other problems: Microsoft admits that "offline files" in Windows XP cannot cope with "large files" and therefore does not cache/sync them at all which would mean those files become unavailable if the connection drop In Windows 7 the feature is only available in the Professional/Ultimate/Enterprise editions. Summary Unless there is another fault-tolerant DFS that supports Windows natively I assume that stacking a HA Samba cluster on top of something like GlusterFS/Lustre/whatnot is the only option, but I hope that I am wrong here. How do other companies allow fault-tolerant network access to redundant storage in a heterogeneous environment with Windows?

    Read the article

  • moving from Exchange 2003 to Exchange 2010

    - by pcampbell
    Consider a small-medium business' deployment of Exchange 2003. The question is around migrating to Exchange 2010. Here's a bit about the landscape: Current state is 50-100 users/mailboxes with the majority using Outlook 2007 OWA enabled desktop users are NOT running in Cached Exchange Mode laptops users ARE running in Cached Exchange Mode a single Exchange server with modest or reasonable specs for the day (3gz, multi-core, 4gb, Win 2003 32-bit) Questions Do you have any suggestions for the admin team regarding the upgrade path/steps from Exchange 2003 to 2010? Considering the requirement of a 64 bit OS, consider a new separate machine as ready to go with Win 2008. Have I missed any details? Where might virtualization help in this project? Any lessons learned in previous upgrades (2007 or 2010) would be appreciated!

    Read the article

  • Radeon 5850 Why am I not getting 3 monitors up as a choice ??

    - by Jan
    Ive just bought the top end ATI Radeon card with 2 normal monitor ports and a HDMI. The idea was to continue using my dual screen setup as always and to use the last plug, the HDMI on my TV. I got a new 52 inch HD TV with all the necessary bits. This should work fine. But.. in Display Properties I still get only my 2 monitors up as options. Not the Digital TV. When I unplug 1 monitor and restart the computer, I get the TV and the other monitor. But never all 3 at the same time. Why is this ? Where can I go to tell it that I need all 3 screens at the same time. Also I get a message saying my gfx card also gives sound through the HDMI cable.. But the TV tells me its recieving a sound format that it does not understand. Any ideas on that too while were at it ?

    Read the article

  • Word list sources

    - by warren
    I am looking for a source of nouns, adverbs, adjectives, and verbs in several languages. I'd like the lists to already be split apart, and not have to go through the OED (and non-English equivalents) by hand re-creating said lists. I don't really care about definitions, and I understand some words can be multiple parts of speech - that's fine - words like "many" could be a noun or adjective, and can appear in both lists. Does anyone here know of such a source? If not, might someone be able to point me in the right direction?

    Read the article

  • Splitting Ethernet cable

    - by Nathan
    I am looking for the best way to network my small business. I have an office where my router/modem are located, and then 2 more computers in another room. I can route Ethernet to this room with about 75 ft of cabling, and I would like to do this because my computers on Ethernet are getting 3x the speed as computers on wireless. Is there any way I can use just one Ethernet cable to span the 75 feet, and then maybe 2 smaller ones to go to the two computers from there? If there's a simple way to do this, maybe with a connector that would be ideal, as opposed to buying over 150 ft of cabling. Thanks!

    Read the article

  • XBMC DVB-T and Played Video filling screen

    - by Tubs
    I have a small PC running as an XBMC attached to a Samsung le37m87bd. The PC isn't powerful enough to output a full HD 1080 image at 1920x1080 which is the TVs native resolution (about every 1/2 a second things go extra fast, I assume skipping frames), so I want to reduce this. Annoyingly, the TV does not support any other widescreen resolutions. (720 etc) If I use a a 4:3 image, the TV stretches it to be the 16:9, however, all 16:9 content is stretched sideways as XBMC is sending out a 4:3 image with a 16:9 image inside. Is there anyway I can force XBMC to compensate for this, ie stretch vertically so that black bars are removed, but not stretched horizontally?

    Read the article

  • Snow Leopard and NAS share

    - by Andreas
    Hi, I bought a BlackArmor 220 from Seagate. When I start Snow Leopard the NAS shows up with its full name under shares. The problem is that when I click the NAS only a folder BA_Backup shows up, all the other folders, I created, don't. When I click the NAS, it says I'm connected as "guest". I mean I can connect by "Go to server" and additionally add the folders to the user startup objects, so they show up when I start my Mac, but when the NAS already shows up, why can't I configure it? Where can I change this? Thanks! Cheers, Andi

    Read the article

  • Premature end of script headers

    - by Tony
    I often get a "premature end of script headers" error in my apache log which results to an internal 500 error. I understand what the error message means - that my application did not give the browser the headers it needs (and maybe nothing at all), but the odd thing is that this does not happen all the time. It actually usually happens the first few times I go to my website after a deploy. Could this be a memory issue? Does anyone know how to trouble shoot this? My apache log isn't really telling me anything. I am running a ruby site using the rails framework on ubuntu hardy. thank you!

    Read the article

  • How can I tell a fresh ubuntu 9.0.4 install to boot from an external Mac OS X usb drive?

    - by Richard Dyce
    I have a reformatted Mac mini containing only a fresh install of Ubuntu 9.0.4. The Apple boot loader as been removed, so I can't hold down option, or the c key to get it to reboot form the install DVD. I do have an external drive (with both a usb and firewire interface available) with a bootable cloned OS X system on it. What steps do I need to go through to in order get grub to see the external drive, and to boot from it so that I can reformat the internal disk as an OS X drive?

    Read the article

  • Understanding an Application based on the OS interaction with a Hypervisor

    - by Dewy
    Following I will ask a few specific questions but I will set the stage first. My goal is to monitor Applications in a very odd place - between the OS and a Hypervisor. If you have comments about this probably unachievable goal please do educate me. One good advice or link can save me days of work. Now to my current attempt. I installed on WinXP a VirtualBox (being open-source) and got a guest OS of latest Ubuntu running within. Where should I go next? Can I set the logs to show all memory/CPU/disk instructions of the guest OS? Thanks, Dewy

    Read the article

  • Why can't I register a domain name that a registrar owns?

    - by barfoon
    Hey everyone, I am interested in purchasing a .ca domain name. It was taken, and when I did a WHOIS it showed me that GoDaddy (a registrar) has control of it, however when I go to GoDaddy to purchase it, I have no such option. I have always wondered - why do registrars do this? I've come across this situation before with other domains as well. Is there any way to actually purchase a domain like this for yourself? If anyone can shed some light on this issue, I'd greatly appreciate it. Cheers,

    Read the article

  • is it really necessary to run Apache as a front-end to Glassfish/JBoss/Tomcat?

    - by Caffeine Coma
    I'm primarily a Java developer, and I come to you with a question that straddles the divide between developers and sysadmins. Years ago, when it was a novel thing to run Tomcat as an app server, it was customary to front it with Apache. As I understand it, this was done because: Java was considered "slow", and it was helpful to have Apache serve static content directly. Tomcat couldn't listen to ports 80/443 unless run as root, which was dangerous. Java is no longer considered slow, and I doubt adding Apache to the mix will actually help speed things up. As for the ports issue, there are probably simpler ways to connect app servers to ports 80/443 these days. So my question is- is there really any benefit to fronting Java Webapps with Apache these days? If so, is Apache still the way to go? Should I look at Nginx? Instead of Tomcat I'm using Glassfish, if that matters.

    Read the article

  • Can two users both control a third machine simultaneously using Synergy?

    - by Reason
    I've been a Synergy user for some time now, as I use a PC on the left side of my Mac. My girlfriend and I both have our desks on each side of the other, and we'd like to know if it were possible for the both of us to control the PC in the middle, with our own separate mouse & keyboards. Here's a crude drawing of our setup (1) her pc (2) my pc (3) my mac Currently, 3 is running a synergy server, and 2 is running the client. But like I said, I'm wondering if there's a way for 1 & 3 to both control 2 with their own mouse and keyboard. I'd ~love~ to have it set up where we could go even farther, and have both of our mice & keyboards able to control all 3 computers at the same time, for moments when we need to click or press keys for each other. But that seems a little too much to ask! Any thoughts?

    Read the article

  • Which Message Queue should I choose (must run on Linux)

    - by MHS
    There are many open source Message queues for Linux, and I need some help deciding what I should go for. My problem is simple - I get sent a list of files that needs to be processed. Each job can't be split up, but they are self contained and can be spread to multiple computers. I'm thinking of solving this using a message queue. Multiple clients send a message to a central queue. Each queue has a number of subscribers that will take jobs from that queue when they have finished processing the current job. Ideally it should have the following qualities Message queue must be able to store unprocessed messages in case of a shutdown/reboot A job can only be processed by a single subscriber (don't want duplicate jobs) The subscribers should be able to send jobs of their own, that will be processed by a different set of subscribers. Can anyone suggest a simple to use message queue?

    Read the article

  • Slowdown after repeated hibernation in XP SP3

    - by John
    I have a Dell laptop w/dual core and 4 gigs of memory. Recently my machine has started to go off in never never land for 5 to 20 minutes at a time with a lot of hard drive churning. Eventually it comes back and everything operates properly. If I manage to have task manager up CPU usage shows no more then 3 to 5% during this process. I am on a corporate network and have very good virus protection in place. The problem seems to have increased in frequency since I let an update happen to IE8 and/or I started using hibernate more frequently. Any ideas or suggestions? TIA J

    Read the article

  • Configuring Apache for multiple clients

    - by Chris_K
    Last week I had a question here about suexec / suphp but I tried to accomplish too much. I'm going to narrow the scope a bit and try again. I'd like to configure a LAMP server to host multiple clients. I'd like it to seem (from the client's viewpoint) just like any other shared hosting environment. Web sites in their home directory, no need to muck around with file ownerships to get pages served, etc. It would seem that a configuration that involves suexec and suphp is the way to go(?) I'm specifically looking for a current/modern guide on how to accomplish this (I'll be using CentOS if it matters) and I'm afraid I need more than a link to Apache docs. Are there any good How-To's out there? The few I've found have been pretty out of date, but it is quite possible my search was weak.

    Read the article

  • Dell Vostro 1000 w/fresh XP Pro installation running very slow

    - by ssvarc
    I'm working on a Dell Vostro 1000 laptop. It was running slowly and hanging, antivirus scans turned up nothing and the same with the Dell diagnostics (DFT hung - but Dell's diagnostics gave the HD a pass) so I imaged the HD and installed XP Pro on it. The laptop is still slow! (It takes a noticeable delay to enter words into the Google search box in IE and the computer hangs every time it enters the XP screensaver.) What are the likely possibilities at this point and how would I go about checking for them?

    Read the article

  • Sync Gmail, Google Contacts, Google Calendar with Microsoft Exchange

    - by Steve Dolan
    At my work we only use Microsoft Exchange. As I hate Outlook and much prefer Google's services, I'd like to be able to sync my email, calendar, and contacts to a Gmail account. It looks like Google shut down their Google Sync service for Gmail accounts earlier this year: http://support.google.com/a/bin/answer.py?hl=en&answer=2716936. They are recommending IMAP, CalDAV, and CardDAV. I'm having trouble even setting up IMAP to work with Exchange. Is this the best way to go or is there a better solution?

    Read the article

  • SQL Server: Is it possible to prevent SQL Agent from failing a step on error?

    - by franklinkj
    I have a stored procedure that runs custom backups for around 60 SQL servers (mixes 2000 through 2008R2). Occasionally, due to issues outside of my control (backup device inaccessible, network error, etc.) an individual backup on one or two databases will fail. This causes this entire step to fail, which means any subsequent backup commands are not executed and half of the databases on a given server may not be backed up. On the 2005+ boxes I am using TRY/CATCH blocks to manage these problems and continue backing up the remaining databases. On a 2000 server however, for example, I have no way to prevent this error from failing the entire step: Msg 3201, Level 16, State 1, Line 1 Cannot open backup device 'db-diff(\PATH\DB-DIFF-03-16-2010.DIF)'. Operating system error 5(Access is denied.). Msg 3013, Level 16, State 1, Line 1 BACKUP DATABASE is terminating abnormally. I am simply asking if anything like this is possible in SQL 2000 or if I need to go in a completely different direction.

    Read the article

  • 5v PCI to PCI-X or PCIe adapter?

    - by SiegeX
    We unfortunately have a very expensive ($10K) full-length 5 volt PCI card that we would like to use in the same system as another expensive PCI-X card. As luck would have it, it seems that PCI-X is not backwards compatible with 5v PCI cards. It would be a real shame to have to order a whole new server just to accommodate these two cards together. Does there exist any internal converter/adapter that will allow one to place a full-length 5v PCI card into either a PCI-X or PCIe slot? I've found an external expansion box that suits our needs but it's 1) External and 2) $1100. The only internal adapters I've been able to find go from Low-profile PCI - PCIe; nothing that seems to support full-height, full-length PCI cards.

    Read the article

  • How to switch users without entering password

    - by torbengb
    I'm a newbie running Ubuntu 9.10. I have two users (wife and me), and each user's screensaver is set to lock so that on wakeup, we get to choose which user's desktop to go to. However, Ubuntu requires a password, so this is pretty tedious. I'd like to switch users without entering any password. I know about this trick that works for the boot login, but it doesn't deal with multiple users. Is it possible to set empty passwords for users in Ubuntu, or skip the password in other ways? (I'm expecting real Linux users to suggest that passwordless users must not get any rights and there be an admin user with a strong password. Yes, you're right. But that's not what this question is about. Thanks.)

    Read the article

  • Getting jerky backgrounds on Win7 on iMac 27'' 2560x1440

    - by JohnIdol
    I installed Win7 on bootcamp on my new iMac 27'' (ATI videocard) and everything was good until recently I noticed that the default win7 background (then one on the background on login) looked jerky. When I say jerky I mean the kind of jerky you get if you can't display enough colours, and instead of nice fading shades you just get stripes and jerky patterns. I am on native resolution but even if I go down to 1920x1080 I get the same. This might have happened after a firmware update but as I don't use windows very often I am not too sure it's what caused it. Oh, and when I am playing games everything looks OK (as in not jerky!). Any help appreciated!

    Read the article

  • Is there any website editor (html) for Firefox

    - by blade19899
    There are a lot of webdeveloper addons available, for Firefox, like firebug. But is there a html editor addon for Firefox. Which I can edit my website. Viewing the edits real-time (using localhost or via ftp?) and saving the edits(CTRL+S) I am currenlty using firebug and it would save a lot of time if I could save the (local) .html file using a firebug like addon!(or firebug itself?) I Found a firebug addon to save .css files... now just need one to save html, and I am good to go!

    Read the article

< Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >