Search Results

Search found 3168 results on 127 pages for 'directories'.

Page 87/127 | < Previous Page | 83 84 85 86 87 88 89 90 91 92 93 94  | Next Page >

  • Sharing user controls acrros multiple websites

    - by Jasper
    Hello, I am currently working on a project which involves three different websites with a lot of common functionality. At the moment the common functionality is placed in a different website full of user controls. The problem is sharing the user controls across the multiple websites. Looking around on SO and other websites, the only solution seems the be using virtual directories. As this is a workable solution (we us this at the moment) it doesn't seem as a "clean" solution. Which "best practices" exist on sharing common functionality (including the GUI/HTML) between different site? Is it (for example) possible to create a single Web Application project and deploy subdirectories (each with their own web.config) to different production environments?

    Read the article

  • ghc can't find my cabal installed packages

    - by nont
    I've installed ghc 6.12.3, and then the Haskell Platform. I'm trying to compile a test program: $ ghc test.hs test.hs:3:0: Failed to load interface for `Bindings': Use -v to see a list of the files searched for. so, naturally, I do cabal install Bindings Which works fine, and places the package in ~/.cabal/lib/bindings-0.1.2 The problem is, that when I go to compile again with ghc, it still doesn't find the package I've installed with cabal. compiling in verbose mode gives: ghc -v test.hs Using binary package database: /home/ludflu/ghc/lib/ghc-6.12.3/package.conf.d/package.cache Using binary package database: /home/ludflu/.ghc/x86_64-linux 6.12.3/package.conf.d/package.cache As suggested by another stackoverflow user, I tried: ghc-pkg describe rts > rts.pkg vi rts.pkg # add the /home/ludflu/.cabal/lib to `library-dirs` field ghc-pkg update rts.pkg But to no avail. How to I add the .cabal to the list of package directories to search? Thank you!

    Read the article

  • Why is my launcher icon smaller when placed on the Android desktop?

    - by Richard
    I have an application built for 1.5 and I am adding higher resolution drawables to support hdpi devices in 1.6 and above. My original application icon is 48x48. I created a second launcher icon that is 72x72 and placed it in a res/drawable-hdpi/ directory. When I install the application on my Nexus One the icon looks good when I view it in the application menu. But when I place it on the desktop it appears small relative to other icons, my guess is I'm seeing the 48x48 version with no scaling. These are my two resource directories: res/drawable/appicon.png res/drawable-hdpi/appicon.png Do I need to do something differently either with my directory structure or my manifest file to get things to display correctly? Thanks!

    Read the article

  • Would you use Code Bubbles?

    - by Paulo Santos
    I've read this question mentioning Code Bubbles and I've watched their video presentation. The video is impressive, and does seem a little bit futuristic, but apparently it's somewhat real. But that kept me thinking... Would a developer really use such tool? We, as developers, are used to deal with code files, organizing them in directories, in one way or another, some common IDE (for those language that has them). It would be a great leap to use something like Code Bubbles, as they propose. I, personally, am not sure if I could work in such environment... although I think I would just need some adjusting... but I really don't see my mind working out the kinks of it. What are your thoughts on this?

    Read the article

  • Best practices for using Hg with Grails?

    - by leeand00
    What should I check in/not check in? Since many of the files are sometimes auto-generated I'm not entirely sure how to handle this using version control...does it have something to do with tags? For instance in ANT, I know not to check-in my target/bin directories...but Grails adds another level of confusion to this...since some of code is generated and some of it is not. (It may become clearer as I go...but it seems to be that there needs to be some way of being able to tell what was just generated and what was modified by a developer so that it needs to be placed in version control)

    Read the article

  • Attachment_fu: can't disable :partition option

    - by Nathan Long
    I'm trying to use the Attachment_Fu plugin in a Rails project, and want to customize the paths where uploaded files are saved. The documentation shows this option: :partition # Whether to partiton files in directories like /0000/0001/image.jpg. Default is true. (The 0001 part is an ID from a table.) I don't want that, so I set the partition option to false, like so: class Photo < ActiveRecord::Base has_attachment :content_type => :image, :storage => :file_system, :max_size => 500.kilobytes, :resize_to => '320x200', :thumbnails => {:thumb => '100x100>' }, :partition => false validates_as_attachment end ...but the :partition => false option has no effect. Has anybody else encountered this problem? How did you fix it?

    Read the article

  • Setting Scala "Platform" for NetBeans 6.8 on Ubuntu Lucid Lynx

    - by David
    I'm trying to use NetBeans 6.8 with Scala, and it can't find the "Scala Platform" (whatever that is supposed to be). I'm using Ubuntu Lucid Lynx (fully updated). The libraries are in /usr/share/java, the binaries in /usr/bin, the docs in /usr/share/doc/scala-doc/, and the sources are uninstalled. I think that NetBeans is looking for Scala to be in one single directory (with bin, lib, etc.). I created /usr/share/scala and placed links to the other directories, but it still didn't quite make it. Can someone straighten me out? (And why is it a "platform"? More needless jargon...)

    Read the article

  • What's wrong with my svn:ignore pattern?

    - by boris callens
    I have the pattern svn:ignore datasheets/*/*.pdf It is supposed to ignore all pdfs that are at an arbitrary depth under multiple "datasheet" directories under the current root folder. As an example: say I have a dir structure like this Websites -web1 -dataSheets -AT -ignore.pdf -BE -NL -ignore.pdf -FR -ignore.pdf -ignore2.pdf -licenseAgreements -important.pdf -web2 -datasheets -etc In this example the pattern needs to ignore all the ignore.pdfs without ingoring the important.pdf too. The shown pattern still includes all my pdf files. I know there are a bunch of similar questions, but none of them seem to tackle the problem with the various hierarchy levels.

    Read the article

  • eclispe workspace backup

    - by MarcoS
    What do I lose if I skip the .metadata/ directory when doing the back-up of my eclipse workspace? (is there some documentation describing what eclipse stores in this directory)? I've noticed that it changes very often (essentially every time that I use eclipse (galileo). I've seen this question, but I'm not interested in doing a back-up of plug-ins and settings (also because I'm not sure that they would work properly when restored after a re-installation of my PC or on a new PC). I'm just interested in doing a back-up of my projects (source code, libraries, possible data, .svn and .git directories). So, can I safely ignore the .metadata/ directory?

    Read the article

  • SmartGWT Server File Browser

    - by jluzwick
    Hi All: I have been looking for a SmartGWT example that would show me how to build a File Browser widget that takes the files from the local server's root directory. The user would be shown the files through the browser which they could then select to perform some processing operations. So far I have thought of using SmartGWT's Tree-Data Binding-Load from Local Data Widget and then grabbing a list of the directories using: new File("\").listFiles(); My Question is: Is there a better way to do this? Has someone already thought of this and has an example of their code that I can see? PS: I'm fairly new to GWT and Web Services but fairly competent with Java. If you believe there is a better way to do this (while still doing this through the web and not using Applets, please tell me). Thanks

    Read the article

  • fubar'd a ruby installation on ubuntu, need helping purging it for reinstall

    - by tipu
    i originally followed these steps to get a RoR environmentgoing: https://help.ubuntu.com/community/RubyOnRails that worked. then i put 1.9.2 on here. then i decided i want the original setup, did a "whereis ruby" and deleted those directories manually. that fubar'd my ruby installation. i can't use the repository to what i originally had. how can i purge ruby from completely and start fresh? right now the specific error is this: /usr/lib/ruby/1.8/rubygems.rb:11:in `require': no such file to load -- thread (LoadError) from /usr/lib/ruby/1.8/rubygems.rb:11 from /usr/bin/gem1.8:8:in `require' from /usr/bin/gem1.8:8 tipu@tipu_ubuntu:~$

    Read the article

  • A RecursiveParentChildIterator -- like the RecursiveDirectoryIterator

    - by Stephen J. Fuhry
    There are tons of examples of using the RecursiveIterator to flatten a tree structure.. but what about using it to explode a tree structure? Is there an elegant way to use this, or some other SPL library to recursively build a tree (read: turn a flat array into array of arbitrary depth) given a table like this: SELECT id, parent_id, name FROM my_tree EDIT: You know how you can do this with Directories? $it = new RecursiveDirectoryIterator("/var/www/images"); foreach(new RecursiveIteratorIterator($it) as $file) { echo $file . PHP_EOL; } .. What if you could do something like this: $it = new RecursiveParentChildIterator($result_array); foreach(new RecursiveIteratorIterator($it) as $group) { echo $group->name . PHP_EOL; // this would contain all of the children of this group, recursively $children = $group->getChildren(); } :END EDIT

    Read the article

  • how sort recursively by maximum fileze and counts files type?

    - by user599395
    Hello! I'm beginner in bash programming. I want to display head -n $1 results of sorting files by size in /etc/*. The problem is that at final search, I must know how many directories and files has processed. I compose following code: #!/bash/bin let countF=0; let countD=0; for file in $(du -sk /etc/* |sort +0n | head $1); do if [ -f "file" ] then echo $file; let countF=countF+1; else if [ -d "file" ] then let countD=countD+1; fi done echo $countF echo $countD I have errors at execution. How use find with du, because I must search recursively?

    Read the article

  • How to parse a tar file in C++

    - by Brendan Long
    What I want to do is download a .tar file with multiple directories with 2 files each. The problem is I can't find a way to read the tar file without actually extracting the files (using tar). The perfect solution would be something like: #include <easytar> Tarfile tar("somefile.tar"); std::string currentFile, currentFileName; for(int i=0; i<tar.size(); i++){ file = tar.getFileText(i); currentFileName = tar.getFileName(i); // do stuff with it } I'm probably going to have to write this myself, but any ideas would be appreciated..

    Read the article

  • What does the 'X' in .aspx, docx, xlsx, etc... represent?

    - by Serapth
    It's one of those things you just take for granted until one day someone asks you and you realize you can't answer it. Much like for years I never questioned the use of 1033 directories in Microsoft products for years until one day, someone asked me about it. Around the release of .NET and Office 2007, Microsoft added an x to basically all of their extensions and I frankly took it as representing XML, but that simply doesn't make sense with .aspx. So, I realize this is a very non technical question, but now that the question has been asked of me and my googling hasn't given me an answer, can anyone tell me with authority what the X represents? Is it extended? Xml? Or is there no meaning behind it?

    Read the article

  • Directory file size calculation - how to make it faster?

    - by Xinxua
    Using C#, I am finding the total size of a directory. The logic is this way : Get the files inside the folder. Sum up the total size. Find if there are sub directories. Then do a recursive search. I tried one another way to do this too : Using FSO (obj.GetFolder(path).Size). There's not much of difference in time in both these approaches. Now the problem is, I have tens of thousands of files in a particular folder and its taking like atleast 2 minute to find the folder size. Also, if I run the program again, it happens very quickly (5 secs). I think the windows is caching the file sizes. Is there any way I can bring down the time taken when I run the program first time??

    Read the article

  • TeamCity Multi-Part Build - How to checkout the code just once

    - by Jeff D
    I am trying to create 1 package with multiple build configurations. The first will checkout the code, build it (Solution File configuration), and run nunit tests. If that succeeds, another will then build in release mode. If that succeeds, a final script witll package up the output, and mark it as an artifact. The problem I'm having is that I don't know how to tell TeamCity not to create new directories for each step, and as a result, the steps are failing. Is there a setting for this? It seems like the dependencies tab would be an appropriate place to look, but I don't seem to understand the instructions, and my tinkering so far has been fruitless.

    Read the article

  • Recursive compilation using gcc

    - by curiousexplorer
    I am using the gcc compiler. My project source tree looks like somewhat like this test$~: tree . . |-- folder | |-- hello.cpp | `-- hello.h `-- main.cpp 1 directory, 3 files test$~: The file main.cpp contains the main() function and all the functions invoked by main.cpp lie in the directory named folder So far in all my little projects I never had to put some source code under a sub-directory. What I am looking for, in short, is some gcc command for recursive compilation in sub-directories and their subdirectories and so on... This command should be invoked from the home directory of the code project.

    Read the article

  • zsh behaves weird in screen

    - by Nils Riedemann
    Hi there, I have zsh set up as my default shell. It works fine as long as i am not within a screen. When i start screen it looks as if some dotfiles are not loaded. For example my $PATH isn't set correctly and some directories are missing. I'm not quite sure where to start looking. Since all is fine, as long as i'm not in a screen. My dotfiles can be viewed on github. I also use oh-my-zsh — as you will notice.

    Read the article

  • Reading into directory, unix shell script

    - by dubya
    I am trying to analyze the files/directories inside of a directory using a shell script, for example if the file is readable, if it is a file, if it is a directory, etc. My script is set up to take a directory as input. so I would type 'file.sh directoryname'. However, when I create a for loop to analyze the files, it analyzes the files in my current working directory rather than the specified directory name. This is my broken code: file=$1 set mypath = $file for file in $mypath * do if [ -d $file ] dirCount=`expr $dirCount + 1` fi done Why does this read the working directory instead of the specified directory? Any help is appreciated. Thanks

    Read the article

  • Apache security for multi-user development web server.

    - by mrmartinblue
    I've been searching and reading through documents all morning and understand that I need to use some combination of chown and probably 'jailing' to securely give programmers access to directories on my centos webserver. Here's the situation: I have an apache web server that has any number of virtual sites located in /var/www/site1 /var/www/site2 etc.. I have different developers that need full access both ssh and vsFTP to only the site they are working on. What is the best way to create and maintain security in this scenario. My thought would be to create a new user for each coder, jail that user to the website directory they are allowed to work in, add their user to a group and set the webroot's owner to that group. Any thoughts? Good, bad, ugly? Thanks!

    Read the article

  • Eclipse+Subversive- Multiple src paths in a single project each from a separate parts of a single re

    - by guinaut
    In a nutshell, I would like to have a single eclipse project with multiple src directories each managed by subversive (or subclipse). If the code base were small, I would use several projects. However, the code base is not small (50+ src paths). Does anyone know how to do this with the Subversion plugins for eclipse? All of the src paths derive from modules that in turn have a common point node in the repository. However, the modules are versioned, so I can’t checkout all of them. It looks as though Subclipse and Subversive do not support multiple src paths. Can anyone confirm this? Solutions?

    Read the article

  • In a Maven project, what are reasons for either a nested or a flat directory layout?

    - by Hanno Fietz
    As my Maven project grows, I'm trying to stay on top of the project structure. So far, I have a nested directory layout with 2-3 levels, where there's a POM on each level with module entries corresponding to the directories at that level. POM inheritance (parent property) does not necessarily follow this, and is not relevant for the purpose of this question. Now, while the nested structure seems pretty natural to Maven, and it's nice and clean as long as you are on one particular level, I'm starting to get confused by what I look at in my IDE (Eclipse and IntelliJ IDEA). I had a look at the Apache Felix sources, and they have a pretty complex project in what seems to be a flat directory structure, so I'm wondering if this would be a better way to go. What are some pros and cons for either approach that you have experienced in practice? Note that this question (which I found meanwhile) seems to be very similar. I'll leave it to the community to decide whether this should be closed as a duplicate.

    Read the article

  • How to generate a report of IIS settings for easy perusal?

    - by Bryan
    Hello, I'm looking for a way to (hopefully) create a text file which lists all the settings in IIS for.. Virtual Directories Web sites Which framework is used on a certain website/directory directory of hosted files etc. Basically I want to do some investigation on some of our servers to figure out where certain projects are located, without digging through right clicking and looking for the directory name manually, etc, for every domain we host. The reasoning is that I often need to find access to files/projects I haven't worked on before, but historically, we don't have a strong naming scheme, so you can't just look where something "logically" would be - so, generating a list would be very helpful. Something like this would be awesome, but I'm looking for any tips at all Domainname1.com framework: ASP.NET 1.1 directory: c:\inetpub\wwwroot\domainname1.com Applications hosted at this domain: etc, etc. Plain text, XLS, XML.. anything other than right clicking through the whole list! Thanks!

    Read the article

  • Making the #include square

    - by David
    I'm trying to write a makefile using CC on Solaris 10. [Only the first bit of that really matters, I think]. I have the following rule for foo.o: foo.o: foo.cc common_dependencies.h CC -c foo.cc -I../../common Unfortunately, common_dependencies.h includes all sorts of idiosyncratic trash, in directories not named '.' or '../../common' . Is this just going to have to be a brute force makefile where I ferret out all of the dependency paths? All of the dependencies are somewhere under '../..', but sometimes 1-level down and sometimes 2-levels down.

    Read the article

< Previous Page | 83 84 85 86 87 88 89 90 91 92 93 94  | Next Page >