Search Results

Search found 3168 results on 127 pages for 'directories'.

Page 87/127 | < Previous Page | 83 84 85 86 87 88 89 90 91 92 93 94  | Next Page >

  • svn track brand new code base

    - by Fire Crow
    I'm at a company, we keep recieviing new codebases from a third party vendor. we'd like to track the changes in subversion. is there a way to replace a branch with the new code and track the changes? currently we just delete all files in the branch, and then add the new files and commit. we'd like to track the files, but I havn't found a tool that will easily deal with all the .svn directories found in subfolders. does anyone know a tool that will replace an svn directory with a new branch and create the respective modify add and delete records as if the code base was organically modified?

    Read the article

  • After pip installing uWSGI there's no /etc/uwsgi/ directory - how can I use apps-enabled?

    - by orokusaki
    I've been using apt-get install uwsgi to install uWSGI. Today, I realized I needed a feature that's not available until uWSGI 1.1, and Ubuntu 12.04.1 doesn't have anything after 1.0.x, at least according to my apt-get install uwsgi=1.1 attempt. So, I used: pip install http://projects.unbit.it/downloads/uwsgi-lts.tar.gz After doing so, I get a message prescribing the use of /usr/local/bin/uwsgi to launch the program. I'm not a guru when it comes to compiling from source, but my understanding is that when you do so, nothing will be changed in the /etc/ directory. Is this correct? If not, why don't I have a /etc/uwsgi/ directory and, more specifically, a /etc/uwsgi/apps-enabled/ directory? Should I simply create the directories when installing uWSGI from source? I was hesitant to do so, considering there is no mention of this in the docs (I don't want something that accidentally works, etc.).

    Read the article

  • Setting Scala "Platform" for NetBeans 6.8 on Ubuntu Lucid Lynx

    - by David
    I'm trying to use NetBeans 6.8 with Scala, and it can't find the "Scala Platform" (whatever that is supposed to be). I'm using Ubuntu Lucid Lynx (fully updated). The libraries are in /usr/share/java, the binaries in /usr/bin, the docs in /usr/share/doc/scala-doc/, and the sources are uninstalled. I think that NetBeans is looking for Scala to be in one single directory (with bin, lib, etc.). I created /usr/share/scala and placed links to the other directories, but it still didn't quite make it. Can someone straighten me out? (And why is it a "platform"? More needless jargon...)

    Read the article

  • FileSystemWatcher keeping parent directory

    - by Henry Jackson
    I am using FileSystemWatcher to monitor a folder, and it seems to be preventing the folder's parent from being deleted. For example, I have the file structure: C:\Root\FolderToWatch\... with the FileSystemWatcher targeting FolderToWatch. While my program is running, if I go to Windows Explorer and try to delete Root, I get an error "Cannot delete Root: access is denied". However, if I delete FolderToWatch FIRST, I can then delete Root without incident. (Needless to say, if the FileSystemWatcher is not enabled, I have no problem deleting either folder.) What gives? Why does the FileSystemWatcher hang onto it's target's parent like that? How (if possible) can I stop this behavior? I would like the user to be able to freely move or delete directories in Windows Explorer.

    Read the article

  • ghc can't find my cabal installed packages

    - by nont
    I've installed ghc 6.12.3, and then the Haskell Platform. I'm trying to compile a test program: $ ghc test.hs test.hs:3:0: Failed to load interface for `Bindings': Use -v to see a list of the files searched for. so, naturally, I do cabal install Bindings Which works fine, and places the package in ~/.cabal/lib/bindings-0.1.2 The problem is, that when I go to compile again with ghc, it still doesn't find the package I've installed with cabal. compiling in verbose mode gives: ghc -v test.hs Using binary package database: /home/ludflu/ghc/lib/ghc-6.12.3/package.conf.d/package.cache Using binary package database: /home/ludflu/.ghc/x86_64-linux 6.12.3/package.conf.d/package.cache As suggested by another stackoverflow user, I tried: ghc-pkg describe rts > rts.pkg vi rts.pkg # add the /home/ludflu/.cabal/lib to `library-dirs` field ghc-pkg update rts.pkg But to no avail. How to I add the .cabal to the list of package directories to search? Thank you!

    Read the article

  • SmartGWT Server File Browser

    - by jluzwick
    Hi All: I have been looking for a SmartGWT example that would show me how to build a File Browser widget that takes the files from the local server's root directory. The user would be shown the files through the browser which they could then select to perform some processing operations. So far I have thought of using SmartGWT's Tree-Data Binding-Load from Local Data Widget and then grabbing a list of the directories using: new File("\").listFiles(); My Question is: Is there a better way to do this? Has someone already thought of this and has an example of their code that I can see? PS: I'm fairly new to GWT and Web Services but fairly competent with Java. If you believe there is a better way to do this (while still doing this through the web and not using Applets, please tell me). Thanks

    Read the article

  • Best practices for using Hg with Grails?

    - by leeand00
    What should I check in/not check in? Since many of the files are sometimes auto-generated I'm not entirely sure how to handle this using version control...does it have something to do with tags? For instance in ANT, I know not to check-in my target/bin directories...but Grails adds another level of confusion to this...since some of code is generated and some of it is not. (It may become clearer as I go...but it seems to be that there needs to be some way of being able to tell what was just generated and what was modified by a developer so that it needs to be placed in version control)

    Read the article

  • Using Tortoise SVN with C++ in Visual Studio 2008

    - by Dr. Monkey
    I have an online repository with some .h and .cpp files that make up part of a project. I'm trying to check these out and use them in a new project, but am getting errors (C4627 and C1010). All the files have been added to the project (with AddExisting Item...), and the subdirectories that contain these files have been added to the "Additional include directories" of the project. Would I be better off having the entire project tree in the repository? My reason for not doing so is that my colleague and I are working on different parts of the code and so want to use different main methods to test things as we go, and I didn't see any need to be passing around any compiled code etc. since I assumed that given the .h and .cpp files (with the correct settings), visual studio would be able to compile the project. What's the best way to make Visual Studio 2008 and TortoiseSVN work well together (without spending any money)?

    Read the article

  • Sharing user controls acrros multiple websites

    - by Jasper
    Hello, I am currently working on a project which involves three different websites with a lot of common functionality. At the moment the common functionality is placed in a different website full of user controls. The problem is sharing the user controls across the multiple websites. Looking around on SO and other websites, the only solution seems the be using virtual directories. As this is a workable solution (we us this at the moment) it doesn't seem as a "clean" solution. Which "best practices" exist on sharing common functionality (including the GUI/HTML) between different site? Is it (for example) possible to create a single Web Application project and deploy subdirectories (each with their own web.config) to different production environments?

    Read the article

  • TeamCity Multi-Part Build - How to checkout the code just once

    - by Jeff D
    I am trying to create 1 package with multiple build configurations. The first will checkout the code, build it (Solution File configuration), and run nunit tests. If that succeeds, another will then build in release mode. If that succeeds, a final script witll package up the output, and mark it as an artifact. The problem I'm having is that I don't know how to tell TeamCity not to create new directories for each step, and as a result, the steps are failing. Is there a setting for this? It seems like the dependencies tab would be an appropriate place to look, but I don't seem to understand the instructions, and my tinkering so far has been fruitless.

    Read the article

  • What's wrong with my svn:ignore pattern?

    - by boris callens
    I have the pattern svn:ignore datasheets/*/*.pdf It is supposed to ignore all pdfs that are at an arbitrary depth under multiple "datasheet" directories under the current root folder. As an example: say I have a dir structure like this Websites -web1 -dataSheets -AT -ignore.pdf -BE -NL -ignore.pdf -FR -ignore.pdf -ignore2.pdf -licenseAgreements -important.pdf -web2 -datasheets -etc In this example the pattern needs to ignore all the ignore.pdfs without ingoring the important.pdf too. The shown pattern still includes all my pdf files. I know there are a bunch of similar questions, but none of them seem to tackle the problem with the various hierarchy levels.

    Read the article

  • Why is my launcher icon smaller when placed on the Android desktop?

    - by Richard
    I have an application built for 1.5 and I am adding higher resolution drawables to support hdpi devices in 1.6 and above. My original application icon is 48x48. I created a second launcher icon that is 72x72 and placed it in a res/drawable-hdpi/ directory. When I install the application on my Nexus One the icon looks good when I view it in the application menu. But when I place it on the desktop it appears small relative to other icons, my guess is I'm seeing the 48x48 version with no scaling. These are my two resource directories: res/drawable/appicon.png res/drawable-hdpi/appicon.png Do I need to do something differently either with my directory structure or my manifest file to get things to display correctly? Thanks!

    Read the article

  • Would you use Code Bubbles?

    - by Paulo Santos
    I've read this question mentioning Code Bubbles and I've watched their video presentation. The video is impressive, and does seem a little bit futuristic, but apparently it's somewhat real. But that kept me thinking... Would a developer really use such tool? We, as developers, are used to deal with code files, organizing them in directories, in one way or another, some common IDE (for those language that has them). It would be a great leap to use something like Code Bubbles, as they propose. I, personally, am not sure if I could work in such environment... although I think I would just need some adjusting... but I really don't see my mind working out the kinks of it. What are your thoughts on this?

    Read the article

  • How to open files in Java Swing without JFileChooser

    - by ron
    I'm using Java Swing (GUI) and I want to add a button to my project for opening files . I don't like the JFileChooser since it opens a small window for browsing through the files of the directories . Can I use something else , instead of the JFileChooser under Java Swing ? I've tried to use elements of SWT but it didn't work , meaning is the use of the button object and then use it inside the Jframe , but that failed , so I guess SWT and Swing don't mix together? Here is the example of Java Swing with JFileChooser and I'm looking for something like this to put in my JFrame.

    Read the article

  • how sort recursively by maximum fileze and counts files type?

    - by user599395
    Hello! I'm beginner in bash programming. I want to display head -n $1 results of sorting files by size in /etc/*. The problem is that at final search, I must know how many directories and files has processed. I compose following code: #!/bash/bin let countF=0; let countD=0; for file in $(du -sk /etc/* |sort +0n | head $1); do if [ -f "file" ] then echo $file; let countF=countF+1; else if [ -d "file" ] then let countD=countD+1; fi done echo $countF echo $countD I have errors at execution. How use find with du, because I must search recursively?

    Read the article

  • fubar'd a ruby installation on ubuntu, need helping purging it for reinstall

    - by tipu
    i originally followed these steps to get a RoR environmentgoing: https://help.ubuntu.com/community/RubyOnRails that worked. then i put 1.9.2 on here. then i decided i want the original setup, did a "whereis ruby" and deleted those directories manually. that fubar'd my ruby installation. i can't use the repository to what i originally had. how can i purge ruby from completely and start fresh? right now the specific error is this: /usr/lib/ruby/1.8/rubygems.rb:11:in `require': no such file to load -- thread (LoadError) from /usr/lib/ruby/1.8/rubygems.rb:11 from /usr/bin/gem1.8:8:in `require' from /usr/bin/gem1.8:8 tipu@tipu_ubuntu:~$

    Read the article

  • A RecursiveParentChildIterator -- like the RecursiveDirectoryIterator

    - by Stephen J. Fuhry
    There are tons of examples of using the RecursiveIterator to flatten a tree structure.. but what about using it to explode a tree structure? Is there an elegant way to use this, or some other SPL library to recursively build a tree (read: turn a flat array into array of arbitrary depth) given a table like this: SELECT id, parent_id, name FROM my_tree EDIT: You know how you can do this with Directories? $it = new RecursiveDirectoryIterator("/var/www/images"); foreach(new RecursiveIteratorIterator($it) as $file) { echo $file . PHP_EOL; } .. What if you could do something like this: $it = new RecursiveParentChildIterator($result_array); foreach(new RecursiveIteratorIterator($it) as $group) { echo $group->name . PHP_EOL; // this would contain all of the children of this group, recursively $children = $group->getChildren(); } :END EDIT

    Read the article

  • Directory file size calculation - how to make it faster?

    - by Xinxua
    Using C#, I am finding the total size of a directory. The logic is this way : Get the files inside the folder. Sum up the total size. Find if there are sub directories. Then do a recursive search. I tried one another way to do this too : Using FSO (obj.GetFolder(path).Size). There's not much of difference in time in both these approaches. Now the problem is, I have tens of thousands of files in a particular folder and its taking like atleast 2 minute to find the folder size. Also, if I run the program again, it happens very quickly (5 secs). I think the windows is caching the file sizes. Is there any way I can bring down the time taken when I run the program first time??

    Read the article

  • In a Maven project, what are reasons for either a nested or a flat directory layout?

    - by Hanno Fietz
    As my Maven project grows, I'm trying to stay on top of the project structure. So far, I have a nested directory layout with 2-3 levels, where there's a POM on each level with module entries corresponding to the directories at that level. POM inheritance (parent property) does not necessarily follow this, and is not relevant for the purpose of this question. Now, while the nested structure seems pretty natural to Maven, and it's nice and clean as long as you are on one particular level, I'm starting to get confused by what I look at in my IDE (Eclipse and IntelliJ IDEA). I had a look at the Apache Felix sources, and they have a pretty complex project in what seems to be a flat directory structure, so I'm wondering if this would be a better way to go. What are some pros and cons for either approach that you have experienced in practice? Note that this question (which I found meanwhile) seems to be very similar. I'll leave it to the community to decide whether this should be closed as a duplicate.

    Read the article

  • What does the 'X' in .aspx, docx, xlsx, etc... represent?

    - by Serapth
    It's one of those things you just take for granted until one day someone asks you and you realize you can't answer it. Much like for years I never questioned the use of 1033 directories in Microsoft products for years until one day, someone asked me about it. Around the release of .NET and Office 2007, Microsoft added an x to basically all of their extensions and I frankly took it as representing XML, but that simply doesn't make sense with .aspx. So, I realize this is a very non technical question, but now that the question has been asked of me and my googling hasn't given me an answer, can anyone tell me with authority what the X represents? Is it extended? Xml? Or is there no meaning behind it?

    Read the article

  • SVN: authz directory identifiers not support

    - by ledy
    Without using the authz, all the svn users can login and use the repos without issues. However, I would like to limit the access to some directories - not to be readable or writeable for all users. svnserve --version = 1.6.6 I tried both, granting access to users and groups. I also tried it separate, only group or only user access. [groups] admingroup=i_can_access_anything limitedgroup=i_am_limited [/] #*= @admingroup=rw i_can_access_anything=rw [projectX] #i also tried [repository:/projectX] #*= @limitedgroup=rw i_am_limited=rw Trying to access the / or /projectX at the svn fails. = access denied Without the authz, it works properly, but also grants access to other projects that do not belong to the "limited" user group :-/ Do you see what's wrong there? Thx

    Read the article

  • Apache security for multi-user development web server.

    - by mrmartinblue
    I've been searching and reading through documents all morning and understand that I need to use some combination of chown and probably 'jailing' to securely give programmers access to directories on my centos webserver. Here's the situation: I have an apache web server that has any number of virtual sites located in /var/www/site1 /var/www/site2 etc.. I have different developers that need full access both ssh and vsFTP to only the site they are working on. What is the best way to create and maintain security in this scenario. My thought would be to create a new user for each coder, jail that user to the website directory they are allowed to work in, add their user to a group and set the webroot's owner to that group. Any thoughts? Good, bad, ugly? Thanks!

    Read the article

  • Upgrading PEAR in command prompt: "Error:Failed to mkdir C:\..."

    - by noscript
    I've just installed PEAR into WAMP attempting to follow this article. It all seems to have gone okay but when I type "pear upgrade" into the command prompt I get this error: ERROR: failed to mkdir c:\php\pear\docs\Archive_Tar\docs It is looking on the wrong drive (and missing a subdirectory in that path by the looks of it). It should be looking for D:\php\php5.3.13\pear\docs\Archive_Tar\docs When specifying installation directories I chose ones on the D:\ and thats where the folder structure is, the include_path in both php.ini locations have also been updated. I've also reset the WAMP server after making the changes. How do I set it to look / install the package in the right place?

    Read the article

  • Eclipse+Subversive- Multiple src paths in a single project each from a separate parts of a single re

    - by guinaut
    In a nutshell, I would like to have a single eclipse project with multiple src directories each managed by subversive (or subclipse). If the code base were small, I would use several projects. However, the code base is not small (50+ src paths). Does anyone know how to do this with the Subversion plugins for eclipse? All of the src paths derive from modules that in turn have a common point node in the repository. However, the modules are versioned, so I can’t checkout all of them. It looks as though Subclipse and Subversive do not support multiple src paths. Can anyone confirm this? Solutions?

    Read the article

  • Reading into directory, unix shell script

    - by dubya
    I am trying to analyze the files/directories inside of a directory using a shell script, for example if the file is readable, if it is a file, if it is a directory, etc. My script is set up to take a directory as input. so I would type 'file.sh directoryname'. However, when I create a for loop to analyze the files, it analyzes the files in my current working directory rather than the specified directory name. This is my broken code: file=$1 set mypath = $file for file in $mypath * do if [ -d $file ] dirCount=`expr $dirCount + 1` fi done Why does this read the working directory instead of the specified directory? Any help is appreciated. Thanks

    Read the article

< Previous Page | 83 84 85 86 87 88 89 90 91 92 93 94  | Next Page >