Search Results

Search found 14206 results on 569 pages for 'compressed folder'.

Page 169/569 | < Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >

  • Add in the header of the license type is enough to say: "my code is licensed"? (Open-source)

    - by silverfox
    I do not know if this is the correct place to ask this stackexchange. Note: If a moderator can move to the correct place (if I am in the inappropriate site SE) I read on various sites about licenses. I did just put the license type in the header file (in my case the javascript file - open-source). /* * "codeName" "version" * http://officialsite.com/ * * Copyright 2012 "codeName" * Released under the "LICENSE NAME" license * http://officialsite.com/LICENSE NAME */ javascript code ... In the same folder I leave a copy of the license. The listing of the folder looks like this: * codeName.js * LICENSE In the file LICENSE would leave my code uses. What nobody says is if it is enough to say my code is licensed (the case of an open-source). Or is something more required? Sorry for the bad English. Thanks.

    Read the article

  • How can I instruct nautilus to pre-generate thumbnails?

    - by Glutanimate
    I have a large library of PDF documents (papers, lectures, handouts) that I want to be able to quickly navigate through. For that I need thumbnails. At the same time however, I see that the ~/.thumbnails folder is piling up with thumbs I don't really need. Deleting thumbnail junk without removing the important thumbs is impossible. If I were to delete them, I'd have to go to each and every folder with important PDF documents and let the thumbnail cache regenerate. I would love to able to automate this process. Is there any way I can tell nautilus to pre-cache the thumbs for a set of given directories? Note: I did find a set of bash scripts that appear to do this for pictures and videos, but not for any other documents. Maybe someone more experienced with scripting might be able to adjust these for PDF documents or at least point me in the right direction on what I'd have to modify for this to work with PDF documents as well.

    Read the article

  • Ubuntu One not syncing and not providing feedback

    - by Joe
    Firstly, I apologize for not being more specific with my problem, but Ubuntu One is just not providing me with any information. It appears to be working, it states that "syncronization in progress..." but it never actually syncronizes (by never I mean 3 days). When I first selected a folder to sync using Ubuntu One it took on the order of hours to sync over 500MB of files - it uploaded the folder hierarchy first and populated the folders over a few hours. That is not happening at all now. Please let me know if there is a way I can get more information out of Ubuntu One that I can post and hopefully resolve this issue. Thanks, Joe

    Read the article

  • The Best How-To Geek Articles for November 2012

    - by Asian Angel
    Last month we covered topics such as why 64-bit Windows needs a separate “Program Files (x86)” folder, how to uninstall your Windows product key before selling your PC, how to deal with locked files in Windows, and more. Join us as we look back at the best articles for November. HTG Explains: Does Your Android Phone Need an Antivirus? How To Use USB Drives With the Nexus 7 and Other Android Devices Why Does 64-Bit Windows Need a Separate “Program Files (x86)” Folder?

    Read the article

  • How can i fit 2 commands in 1 terminal shortcut

    - by Nicky Bailuc
    10 latest updates and drivers and I need to run a game called unreal tournament, but in terminal it requires 2 commands The first one is to mount into the folder: cd /usr/local/games/ut2004/ and then the second one is to open the actual game: sudo aoss ./ut2004 In one shortcut i can only fit 1 command but both don't fit in is there any way i can turn these 2 commands into one? Perhaps turni9ng on the desktop shortcut already mounted into the folder? Any help would be really appreciated because im getting kinda sick of using the terminal to run it every time.

    Read the article

  • Quick path jumping

    - by Sebastian P.
    I was just at a lecture, where I noticed the lecturer using a command (probably aliased) to jump to a specific folder. Example: ~/code$ j sciproj ~/projects/sciproj2011/$ This looked quite slick, so I started wondering: Is this a standard utility, and if so, what is the name? I have two theories as to how it works: It can both create, delete and jump to aliases directly from the command-line in the style of the example, without having to set up aliases in a configuration file or script or whatnot manually. It searches the home directory for a folder matching the name and jumps to it. The second option seems a bit slow, however, so the first would be preferred.

    Read the article

  • Common reasons for the &lsquo;Sys is undefined&rsquo; error in ASP.NET Ajax applications

      In this blog I will try to summarize the most common reasons for getting the famous 'Sys is undefined' error when running an Ajax enabled web site or application (there are almost one milion results on Google for that phrase). Where does it come from? In every Ajax web pages source you will see a code like this: <script type="text/javascript"> //<![CDATA[ Sys.WebForms.PageRequestManager._initialize('ScriptManager1', document.getElementById('form1')); Sys.WebForms.PageRequestManager.getInstance()._updateControls([], [], [], 90); //]]> </script>   This is the initialization script of the ScriptManager. So, if for some reason the Sys namespace is not available when the code executes you get the Sys is undefined error. Here are the most common reasons and solutions for that problem:   1. The error occurs when you have added a control from RadControls for ASP.NET AJAX, but your application is not configured to use ASP.NET AJAX. For example, in VS 2005 you created a new Blank Site instead of a new Ajax-Enabled Web Site and the Sys is undefined message pops up. To fix it you need to follow the steps described at Configuring ASP.NET Ajax article (check the topic called Adding ASP.NET AJAX Configuration Elements to an Existing Web Site) or simply create the Ajax-Enabled Web Site. You can also check my other blog post on the matter: Visual Studio 2008: Where is the new ASP.NET Ajax-Enabled Web Site template?   2. Authentication - as the website denies access to all pages to unauthorized users, access to the Telerik.Web.UI.WebResource.axd handler is unauthorized (this is the default handler of RadScriptManager). This causes the handler to serve the content of the login page instead of the combined scripts, hence the error. To solve it - add a <location> section to the application configuration file to allow access to Telerik.Web.UI.WebResource.axd to all users, like: <configuration> ... <location path="Telerik.Web.UI.WebResource.axd"> <system.web> <authorization> <allow users="*"/> </authorization> </system.web> </location> ... </configuration>   Note that the access to the standard ScriptResource.axd and WebResource.axd is automatically allowed for all users (authenticated and unauthenticated), so if you use the ScriptManager instead of RadScriptManager - you will not face this problem. The authentication problem does not manifest when you disable script combining or use the CDN. Adding the above configuration section will make it work with RadScriptManagers combined script.   3. The IE6 browser fails to load the compressed script. The problem does not appear in any other browser. There is a well known bug in the older versions of IE6 which lose the first 2,048 bytes of data that are sent back from a Web server that uses HTTP compression. Latest versions of RadScriptManager does not compress the output at all if the client is IE6, but in the previous versions you need to manually disable the output compression to prevent the error. So, if you get the Sys is undefined error in IE6 - update to the latest version of RadControls or simply disable the output compression.   4. Requests to the *.axd files returns Error Code 404 - Not Found. This could  be fixed easily: Check in the IIS management console that the .axd extension (the default HTTP handler extension) is allowed:     Also check if the Verify if file exists checkbox is unchecked (click on the Edit button appearing in the previous screenshot to check). More information can be found in our troubleshooting article and from the ASP.NET QA team blog post   5. The virtual directory in IIS is not marked as Web Application. Converting it to Web Application should fix the problem.   6. Check for the <xhtmlConformance mode="Legacy"/> option in your web.config and remove it. It would be rather rare to become a victim of this exact case, but still have it in mind. Scott Guthrie describes it in more details   In the above points I mentioned several times the terms web resources, javascript output, compressed script. If you want to find out more about these please see the Web Resources Demystified series of my friend and colleague Atanas Korchev   I hope that one of the above solutions will help you get rid of the Sys is undefined error.   Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • function names - "standartised" prefixes

    - by dnsmkl
    Imagine you have such routines /*just do X. Fail if any precondition is not met*/ doX() /*take care of preconditions and then do X*/ takeCareOfPreconditionsCheckIfNeededAtAllAndThenDoX() A little bit more concrete example: /*create directory. Most probably fail with error if any precondition is not met (folder already exists, parent does not exists)*/ createDirectory(path_name) /*take care of preconditions (creates full path till folder if needed, checks if not exists yet) and then creates the directory*/ CheckIfNotExistsYet_CreateDirectory_andFullPathIfNeeded(path_name) How do you name such routines, so it would be clear what does what? I have come to some my own "convetion" like: naiveCreateDirectory, ForceDirectoryExists, ... But I imagine this is very standard situation. Maybe there already exists some norms/convetions for this?

    Read the article

  • Mysql not starting - innodb not found

    - by Rob Guderian
    I have a fresh install of ubuntu 12.04 server edition and mysql server is not starting properly. I did a simple apt-get install apt-get install mysql-server But, it's failing with this error message root@test:~# mysqld 120618 20:57:32 [Warning] The syntax '--log-slow-queries' is deprecated and will be removed in a future release. Please use '--slow-query-log'/'--slow-query-log-file' instead. 120618 20:57:32 [Note] Plugin 'FEDERATED' is disabled. 120618 20:57:32 InnoDB: The InnoDB memory heap is disabled 120618 20:57:32 InnoDB: Mutexes and rw_locks use GCC atomic builtins 120618 20:57:32 InnoDB: Compressed tables use zlib 1.2.3.4 120618 20:57:32 InnoDB: Unrecognized value fdatasync for innodb_flush_method 120618 20:57:32 [ERROR] Plugin 'InnoDB' init function returned error. 120618 20:57:32 [ERROR] Plugin 'InnoDB' registration as a STORAGE ENGINE failed. 120618 20:57:32 [ERROR] Unknown/unsupported storage engine: InnoDB 120618 20:57:32 [ERROR] Aborting I can start the server with the "--skip-innodb --default-storage-engine=myisam" flags, but would like to use innodb. Does anyone know what the issue here is?

    Read the article

  • Ubuntu One not syncing fully

    - by wurlyfan
    I have uploaded several folders of data to Ubuntu One from my desktop computer, over the last few weeks, and I can see that all the contents are there when I look at my account on the web. When I look at my laptop (connected to the same account), one of the first folders I uploaded hasn't downloaded completely. New folders added from either device seem to sync correctly, but this one older folder remains almost empty, even though the control panel says file syncing is up-to-date. I have plenty of space available. Stopping and restarting the sync daemon and rebooting the laptop are both ineffective. What can I do to make this folder sync fully? I don't want to risk losing the data (which now exists only in Ubuntu One), and I don't have a lot of broadband data to play with. I've seen several bugs relating to this sort of issue but they're all quite old and apparently fixed, while this is happening on new 13.04 installations on both desktop and laptop.

    Read the article

  • Is adding in the header the license type enough to say: "my code is licensed"?

    - by silverfox
    I read on various sites about licenses. I did just put the license type in the header file (in my case a javascript file, open-source): /* * "codeName" "version" * http://officialsite.com/ * * Copyright 2012 "codeName" * Released under the "LICENSE NAME" license * http://officialsite.com/LICENSE NAME */ javascript code ... In the same folder I leave a copy of the license. The listing of the folder looks like this: * codeName.js * LICENSE In the file LICENSE is the full text of the license my code uses. What I cannot find anywhere that says is this is enough to say my code is licensed (the case of open-source). Is something more required?

    Read the article

  • Lubuntu 14.04 Problem starting lxsession-default-apps

    - by user278179
    I have one problem, I can't execute lxsession-default-apps on Lubuntu 14.04 because I get because said to me "The database is updating, please wait" If I try to run lxsession-default-apps, I get this error: ** Message: utils.vala:30: config_path_directory: /home/USER/.config/lxsession-default-apps ** Message: desktop-files-backend.vala:171: test config_path: /home/USER/.config/lxsession-default-apps/settings.conf ** Message: desktop-files-backend.vala:237: Scanning folder: /usr/share/applications ** Message: desktop-files-backend.vala:278: Start scanning ** Message: desktop-files-backend.vala:257: Scanning folder: /usr/share/app-install/desktop ** Message: desktop-files-backend.vala:278: Start scanning Error: list_files failed: No such file or directory ** Message: desktop-files-backend.vala:333: Finishing scanning ** Message: desktop-files-backend.vala:189: Signal finish scanning with mode: write ** Message: desktop-files-backend.vala:333: Finishing scanning Any help would be appreciated. Thanks. Regards.

    Read the article

  • Subfolder non-www redirect

    - by Zealotry
    Is there a way to redirect a sub folder to no-www? What I use is: RewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} ^www\.(.*)$ [NC] RewriteRule ^(.*)$ http://%1/$1 [R=301,L] but this only redirects the www.example.com to example.com. I would like to redirect: www.example.com/home/ to example.com/home/ www.example.com/home/whatever URL to example.com/home/whatever URL. I have tried this: RewriteEngine on Options +FollowSymlinks -MultiViews RewriteCond %{HTTP_HOST} ^(www\.)?example\.com$ [NC] RewriteRule ^$ http://example.com [R=301,L] RewriteCond %{HTTP_HOST} ^www\.example\.com$ [NC] RewriteCond %{REQUEST_URI} !^/home/ [NC] RewriteRule ^(.+)$ http://example.com/$1 [R=301,L] This does not work, either. I can't really figure it out. Any help appreciated! ANSWER I figured it out and I will post it for others to see, if having the same issue. In the sub folder directory's .htaccess used the following: RewriteCond %{HTTP_HOST} ^(www\.example\.com)?$ RewriteRule ^(.*)$ http://example.com/subfoldername/$1 [R=301,L]

    Read the article

  • Question about metacity issue from previous post

    - by Alex Poulos
    Yesterday I had an issue with my 12.04 - 12.04 Does anyone know how to fix this? all of a sudden - my Terminal shortcut doesnt work and windows only appear in top left corner and some configuration causing an issue with metacity making it crash - I was walked through the problem, and found out that by creating another user account and logging onto it - the problem wasn't present anymore on that account but still was present on my old account. So my question is - was the problem just something in my old Home folder? - also - can most issues that occur like this be solved by making a new account and backing up the "Home" folder?

    Read the article

  • Feed Reader Fix

    - by Geertjan
    In the FeedReader sample (available in the New Projects window), there's this piece of code: private static Feed getFeed(Node node) { InstanceCookie ck = node.getLookup().lookup(InstanceCookie.class); if (ck == null) { throw new IllegalStateException("Bogus file in feeds folder: " + node.getLookup().lookup(FileObject.class)); } try { return (Feed) ck.instanceCreate(); } catch (ClassNotFoundException ex) { Exceptions.printStackTrace(ex); } catch (IOException ex) { Exceptions.printStackTrace(ex); } return null; } Since 7.1, for some reason, the above doesn't work. What does work, and is simpler, is this, instead of the above: private static Feed getFeed(Node node) { Feed f = FileUtil.getConfigObject("RssFeeds/sample.instance", Feed.class); if (f == null) { throw new IllegalStateException("Bogus file in feeds folder: " + node.getLookup().lookup(FileObject.class)); } return f; } So, the code needs to be fixed in the sample.

    Read the article

  • On-the-fly file search in gnome 3?

    - by jan groth
    I recently switched to Gnome 3 (on Onereic 11.10). I was a big fan of gnome-do (on Gnome 2), where I heavily used the "files and folders"-plugin. This allows to configure a list of files and folder to be searched on-the-fly while typing. E.g. starting to type "timesheet" opens "timesheet_2011.ods" from my dropbox account... I can't find anything like this in Gnome 3. I know that there is a list of "recent items", and also the name of bookmarked folders in nautilus is searched, but that's not (exactly) what I'm looking for... So, what I want is this: tell Gnome 3 where my files are (certainly not only in my home folder) have them searched on the fly while I'm typing I think Unity offers something like this - but hey, I'm using Gnome 3 ;-) Anything I missed?

    Read the article

  • Missing feature in Hyper-V from Virtual PC

    - by Kevin Shyr
    One thing I really miss is the ability to create shared folder between host and guest.  Virtual PC does this well, you can create Shared Folder to be used every time, or just this one.  I have read some posts on how to do this.  Some people suggest using ISO Creator to package up the files and mount the image to DVD drive, but what I need is truly a "shared" environment, so I'm currently looking into creating Virtual switch and creating an internal network between the host and guest.  Let's see how that works out. I would have loved to give Virtual SAN Manager a try, but I don't have a local Fibre Channel to set one up. I guess this might be an extension to my original post:  http://geekswithblogs.net/LifeLongTechie/archive/2011/05/05/windows-virtual-pc-vs.-hyper-v-virtual-machines-vs.-windows-virtual.aspx

    Read the article

  • "Failed to mount Windows share" error in Samba

    - by Ranjith R
    This is the situation. There are 3 machines in the office. The Operating systems on them are respectively, Linux mint Ubuntu 12.04 Windows Vista The Ubuntu (#2) machine is supposed to be the common file server between the machines #1 and #3. Machine #2 has two hard disks. One is a 500 GB NTFS empty drive and the other is a 160 GB ext4 drive. My plan is to make the 500 GB as the file sharing disk. When I share a folder like ~/Documents using Nautilus context menu on machine #2, I can access the files easily on both #1 and #3, but when I try to share some folder on 500 GB disk, I get an error on machine #1 that says Failed to mount windows share I do not mind formatting the drive to ext4 if needed, but I am sure that something simple is wrong. EDIT I took @Marty's comment as a hint and used ntfs-config to configure automount of that partition. It is working now. Thanks

    Read the article

  • Term for Production Rollback

    - by trainer
    The firm I work at keeps a copy of the compiled source code on production. So basically on our Production server, we have two folders: ProductionServer\SourceCode\ ProductionServer\SourceCodeRollback\ The Rollback folder always contains one version older code. This is done so that on deployment nights, if something is wrong with the new code, the deployer(who is not familiar with source control software) has the older version to revert back to. My question is: Is there an official computer science term/nomenclature for the Rollback folder? Or is "Rollback" itself the official term?

    Read the article

  • photshop: why low quality jpg saved as high quality increases the file size?

    - by Alex Angelico
    i have a low quality background about 85kb. If I open the file in Photoshop and save it, I have to save this image as 100% quality, the file size increases to 640kb. This doesn't make sense, the image si already compressed, the quality cannot be better than the source, so the 100% quality while saving should produce THE SAME FILE OPENED. The problem is, I have this background and want to add a logo image over it, and then save it. But If I do so, i have to save this with 10% quality, and the logo looks horrible. If I save this to 100% quality, the file size is huge... How can I achive this?

    Read the article

  • php file upload problem [closed]

    - by newcomer
    This code works properly in my localhost. I am using xampp 1.7.3. but when I put it in the live server it shows Possible file upload attack!. 'upload/' is the folder under 'public_html' folder on the server. I can upload files via other script in that directory. <?php $uploaddir = '/upload/';//I used C:/xampp/htdocs/upload/ in localhost. is it correct here? $uploadfile = $uploaddir . basename($_FILES['file_0']['name']); echo '<pre>'; if (move_uploaded_file($_FILES['file_0']['tmp_name'], $uploadfile)) { echo "File is valid, and was successfully uploaded.\\n"; } else { echo "Possible file upload attack!\\n"; } echo 'Here is some more debugging info:'; print_r($_FILES); print "</pre>"; ?>

    Read the article

  • U1 music shows unknown artist, how can I make it recognize m4a files?

    - by bisi
    Hello all, I am in the process of uploading my music library as a backup to U1, but I figured, why not also enjoy Ubuntu 1 music on my iPhone? With a few difficulties to start, the upload is now in progress, but I've noticed that there is a huge percentage of files in the unknown artist folder, and I believe it is all of my m4a files. They play fine, but without any information. Coming from an iTunes background, and having bought the majority of my music on the iTunes store, I wonder how I could make this work, easily? I am on Maverick (afaik), but About Ubuntu shows 11.04. I use Banshee as my music manager, and I monitor my sync using Ubuntu one preferences, ubuntuone-indicator and magicicada. The total file size of my music folder is 38.9GB. Thank you for your help!! And apologies if I couldn't find a thread where this was already covered...

    Read the article

  • Function names - "standardized" prefixes

    - by dnsmkl
    Imagine you have such routines /*just do X. Fail if any precondition is not met*/ doX() /*take care of preconditions and then do X*/ takeCareOfPreconditionsCheckIfNeededAtAllAndThenDoX() A little bit more concrete example: /*create directory. Most probably fail with error if any precondition is not met (folder already exists, parent does not exists)*/ createDirectory(path_name) /*take care of preconditions (creates full path till folder if needed, checks if not exists yet) and then creates the directory*/ CheckIfNotExistsYet_CreateDirectory_andFullPathIfNeeded(path_name) How do you name such routines, so it would be clear what does what? I have come to some my own "convetion" like: naiveCreateDirectory, ForceDirectoryExists, ... But I imagine this is very standard situation. Maybe there already exists some norms/convetions for this?

    Read the article

  • Why are downloads from Canonical Partners repository so slow?

    - by Sabacon
    If I need Sun Java, Adobe Flash Plugin or anything else that comes from Canonical Partners the package downloads are painfully slow even small sized packages like the Flash plugin, to speed things up I have to go here: http://archive.canonical.com/ubuntu/pool/partner/ to find what I want, download the packages with a download manager (which is usually about 20 times faster than the package manager) and then place them in my /var/cache/apt/archives folder I run the package manager afterwards, as long as the right versions of the packages I ask to install are detected in the /var/cache/apt/archives folder they will be installed immediately. I would like to stop doing this, so I am wondering if anyone else has this problem, what could be the cause and if there is a fix. I am located in the Western Caribbean region. I think it would be helpful to note that all other packages coming from the repository I have selected with synaptic download at acceptable speeds.

    Read the article

  • Weird execution of ruby/git executables in Windows [migrated]

    - by Frexuz
    Something strange has happened. I can't run some command line executables in Windows anymore. Steps: Open cmd Run an executable, such as ruby -v or git -h When I do that, a new command prompt opens, running that command (I think, it's too fast to see), and instantly closes again. I've managed to print screen the new command prompt, and it shows that it's running inside this path: C:\Documents and Settings\Administrator\Local Settings\Temp\3582-490 Inside this folder, is the executable I'm tring to run. If I run ruby, then ruby.exe is in there. If I run git, then git.exe is in there. And it's always emptying the folder in between, so there is always just one .exe file

    Read the article

< Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >