Search Results

Search found 34489 results on 1380 pages for 'folder view'.

Page 489/1380 | < Previous Page | 485 486 487 488 489 490 491 492 493 494 495 496  | Next Page >

  • Linux Log Viewer with Web interface

    - by user180039
    I have been asked at work to find a solution to one of our problems. We have several logs that customers need access to, because we don't want to give them direct access to the folder/share we are looking to implement a simple Web based solution that permits customers to login see a list of files they have permissions to and download the file. It would need to be able to setup permissions so User01 can see file01 and file03 and User02 can see file04 and file06, optimally all the files would be under the same folder, so permissions are based on files rather then based on folders. Anyone got any ideas Many Thanks

    Read the article

  • Restore Picasa people tags

    - by Paul
    I have loaded Windows 7 to my laptop. Before doing this I backed up all my pictures using the Picasa backup utility. I then ran a restore on the clean Windows 7 install. I then installed Picasa 3.5 and none of the people tags showed up. I then went and deleted what I thought was the Picasa DB and then tried running the restore again. Now each folder shows up twice in Picasa but only once under the Windows Pictures folder. How do I get rid of the duplicates in Picasa and get my people tags back?

    Read the article

  • Black Screen on pressing back button in libgdx

    - by user26384
    In my game when i touch on advertisements and press back button to return on game the i am getting a black screen. I referred this link. I tried to change IosGraphics.java but the change is not reflected in monotouch project. I did the following : Extracted nightly.zip and opened gdx-backend-iosmonotouch-sources From there I changed IosGraphicsjava. I then made a new jar file gdx-backend-iosmonotouch.jar and replaced it with original jar file in the nightly folder. Compressed all the files from nightly folder in .zip file. Used this .zip file to make a new project throuch gdx-setup-ui.jar. I tried to open my project in monotouch and from com-gdx-backendios.dll i found that the changes in IosGraphics are not being reflected. Am I missing something? How do I solve this? I even tried to open gdx-backend-iosmonotouch-sources.jar with winrar and edit IosGraphics.java and save it. Even this didn't work.

    Read the article

  • Cygwin creates files with special (shared) icons on windows

    - by barjonah
    I use cygwin to transfer files between linux and windows machines. Everytime I transfer a file to a windows machine it adds an extra shared user icon on the file's or folder's icon itself. This also happens if I create a file from cygwin on windows using pretty much any command: echo, vim, nano, cat. This is what the cygwin-created folders (and files) look like. This is what a normal folder (or file) looks like. I'm thinking it has to do with permission, because I'd have to chmod it everytime if I want other applications to access the files or folders on windows. How can I tell cygwin to create regular ol' files just like a user or any program would?

    Read the article

  • Calling a model from a controller from the 404 route [migrated]

    - by IrishRob
    Got a problem here where I can’t seem to load a method from a model after the page has been redirected after encountering a 404. Model name: Category_Model Method name: get_category_menu() In my routes, I’ve updated the 404 over-ride to: $route[‘404_override’] = ‘whoops’; I’ve also got my controller Whoops that reads… <?php class Whoops extends CI_Controller { function index() { $this->load->model('Category_Model'); $data['Categories'] = $this->Category_Model->get_category_menu(); $data['main_content'] = $this->load->view('messages/whoops', null, true); $this->load->view('includes/template', $data); } } So when I navigate to a page that doesn’t exist, I get the following error… Message: Undefined property: Whoops::$Category_Model Filename: controllers/whoops.php I’ve hard coded the loading of the model into the controller here, even though I have it in my autoload, but no luck. Everything else with the site so far works, just this 404 problem. Any pointers would be great, kinda new to CI so go easy on me. Cheers.

    Read the article

  • Installing ethernet controller driver after clean windows XP install

    - by user1488804
    I've been having trouble getting internet access through an ethernet connection since doing a fresh install of windows XP. It appears that I'm missing the driver for the ethernet controller. Looking in 'device manager' under the network adapters section there's an apparently working '1394 Net Adaptor' but under unknown devices there's the question-marked ethernet controller, which I believe is what I want the driver for. I found an old cd that contains the driver for it (Realtek 8139 or some 81xx). There's no setup tool, but only 3 files (a .cat file, .inf file and .sys file). I've tried right-clicking on the .inf file and using 'install' and also tried just copying the .sys file into the system32/drivers folder but neither have got me anywhere after restarting. The ethernet controller is still listed in 'unknown devices'. EDIT: I've also tried updating the driver in the device manager, pointing it at the folder containing the driver, but this hasn't worked either. Any ideas? Thanks

    Read the article

  • Permission denied on network share

    - by Philipp
    i have a Windows 8 host system running a virtual(hyper-v) Debian6 client with an lamp environment. My development environment runs under Windows and I mapped the folder with my php files to a network drive so Apache has access to them.(mount.cifs //pc/share /var/share/) This far no problems - I see my app on windows in the browser. The problem is, I can't write stuff in php to the share folder - everytime i got a permission denied message in my error logs. For testing purpose i tried to change the directory permissions of /var/share with chmod -R 777 /var/share without success. Now Iam a little bit stumped.. has anyone an idea how to solve this?

    Read the article

  • Build 2013&ndash;Keynote Thoughts

    - by D'Arcy Lussier
    Originally posted on: http://geekswithblogs.net/dlussier/archive/2013/06/26/153243.aspxSome thoughts on the Build 2013 keynote. They Listened to Feedback while Keeping to their Plans I am one of the people in the “bring back the start menu” camp. I want my start menu. I *like* my start menu. Microsoft heard that and put it back, fantastic. But they implemented it in a way that still pushes the Windows 8 UI – and I’m actually pretty happy with it. When you hit the Start menu, you get the live-tiles displayed overlaying the desktop. But you can also swipe from the bottom to get the “all-applications” view. This, in essense, is really what those that like the Start Menu want. I believe it was mentioned that you can configure the all-applications view to be the default. They’re Committed to Improving Windows 8 The commitment to rapid deployments Ballmer talked about is crucial to Windows 8’s success. They need to keep it evolving quickly to maintain the interest of users and developers. I think the little improvements they showed are excellent (hands-free mode, multi window docking, better multi-monitor support, new developer controls, etc.). Hardware Vendors are Committed to Windows 8 They showed off a number of new hardware products (Windows 8 and Windows Phone). The Surface’s introduction to the market has done nothing to dissuade their hardware partners. Bing as a Platform is Huge for Developers!!! This was the biggest take-away from the keynote! What the team is doing with Bing not as a search engine but as a developer API is very impressive! I’m going to be diving into this over the rest of Build so watch more blog posts coming on it. Azure, Office 365, and other topics will be covered at tomorrow’s keynote. So far, great kick off to Build. Now on to sessions! D

    Read the article

  • Soft links between samba profile and profile.V2

    - by Alex Rose
    I am currently using a samba server to host windows sessions. For the moment, every machine runs windows XP, but in an upcoming hardware upgrade, I will have new Windows 7 machines installed. The problem is that the user directory changed from XP to 7 (for instance 'My Documents' became 'Documents'). To solve that, samba created a folder called profile.V2, which contains files and folders for Windows 7. Now I would like to link the two profiles folders ('profile' from windows XP, 'profile.V2' from windows 7) so that a user can logon from a XP or 7 machine and still have access to the same files. I tried to create softlinks between folder pairs ('Documents' -­­ 'My documents') and it seems to work. My question is : is it likely to create issues in the future?

    Read the article

  • IT merger - self-sufficient site with domain controller VS thin clients outpost with access to termi

    - by imagodei
    SITUATION: A larger company acquires a smaller one. IT infrastructure has to be merged. There are no immediate plans to change the current size or role of the smaller company - the offices and production remain. It has a Win 2003 SBS domain server, Win 2000 file server, linux server for SVN and internal Wikipedia, 2 or 3 production machines, LTO backup solution. The servers are approx. 5 years old. Cisco network equippment (switches, wireless, ASA). Mail solution is a hosted Exchange. There are approx. 35 desktops and laptops in the company. IT infrastructure unification: There are 2 IT merging proposals. 1.) Replacing old servers, installing Win Server 2008 domain controller, and setting up either subdomain or domain trust to a larger company. File server and other servers remain local and synchronization should be set up to a centralized location in larger company. Similary with the backup - it remains local and if needed it should be replicated to a centralized location. Licensing is managed by smaller company. 2.) All servers are moved to a centralized location in larger company. As many desktop machines as possible are replaced by thin clients. The actual machines are virtualized and hosted by Terminal server at the same central location. Citrix solutions will be used. Only router and site-2-site VPN connection remain at the smaller company. Backup internet line to insure near 100% availability is needed. Licensing is mainly managed by larger company. Only specialized software for PCs that will not be virtualized is managed by smaller company. I'd like to ask you to discuss both solutions a bit. In your opinion, which is better from the operational point of view? Which is more reliable, cheaper in the long run? Easier to manage from the system administrator's point of view? Easier on the budget and easier to maintain from IT department's point of view? Does anybody have any experience with the second option and how does it perform in production environment? Pros and cons of both? Your input will be of great significance to me. Thank you very much!

    Read the article

  • WSUS moved how to reset the database and check the folders?

    - by Matthew Zielonka.co.uk
    I have a bit of a problem with a WSUS box, I backed up the WsusContent folder (about 180 gig) to a second machine, wiped the first machine then realized I don't have the database folder! I can not find any articles / guidance on this, I have re installed WSUS however it does not find the other files which where previsously downloaded (I guess the database being the reason). If I do a reset it will force it to download the ones it has again I guess and does not check what is still to be downloaded? Sigh... love / hate WSUS at the moment with a passion :) Thanks for any help.

    Read the article

  • Auto create folders using Excel field entries

    - by prrao
    I need to generate folders containing certain Excel field entries: Additionally, I need to append a date to the folder name based on that entry So If I have the above Excel table, I'd need the following set of folders to be generated in the same directory: Is this possible using a macro? Also, in case I add another entry how would I generate another folder only for that entry; for example I add another project CWO-1106: Basically, I'd like to keep updating the Excel sheet and create folders only for the new entries Any help would be appreciated. Thanks!

    Read the article

  • Building modules with linux kernel for custom flavor

    - by Chandan Apsangi
    I followed the instructions given in the link: http://blog.avirtualhome.com/how-to-compile-a-new-ubuntu-11-04-natty-kernel/ for building a custom kernel and booting it. Everything works fine, except that when building it, I used the option skipmodule=true (as given in this link), so I guess the modules are not built for this kernel. I'm saying this because after I installed the image (using dpkg -i), when I run 'ls -l' in the /lib/modules/2.6.38-11-i5 folder, all the files/folders except the 'kernel' folder have got updated (Btw, I had built it last time and it had worked fine,this is an update on the same kernel source). So is there some option/command that I missed that exports the modules as well? Thanks for your help.

    Read the article

  • JRE not running correctly on Windows 7 64Bit

    - by dkt91
    An application called DBGL (DosBox Game Launcher) requires JRE in order to run successfully. I had it installed and running on my old Windows 7 32Bit machine without any problems. Today I got my new PC with Windows 64Bit and I cannot get it running anymore. If I install the latest JRE 32Bit version to the \Program Files (x86) folder DBGL says it needs the latest JRE version in order to run and doesn't start. If I install the latest JRE 64Bit version to the \Program Files folder I do not get an error message anymore but the application doesn't start either. Clicking it has literally no effect. Right now I have both JRE versions running parallely and I get the same result I have when only the 64Bit version is installed. Thanks in advance!

    Read the article

  • Syncronization inside the same PC

    - by FGO
    Hi, I work in a company without a server and operating in diferent countries. Because of this we work on a virtual office basis and share files through a virtual server called Drop Box. I'm the controller of this virtual server and find it to be incomplete. We work on levels of permission so I invite access to users for 3 possible levels and find sometimes that I need to have a specific folder inside one mother folder to be inside 2 levels. The only way I can find to make my life easier is to automatically syncronize specific folders between diferent mother folders, but I don't know how to do this in windows 7. Can anyone give me a hint? Thanks

    Read the article

  • Question about modeling with MVC (the pattern, not the MS stuff / non web)

    - by paul
    I'm working on an application in which I'm looking to employ the MVC pattern, but I've come up against a design decision point I could use some help with. My application is going to deal with the design of state-machines. Currently the MVC model holds information about the machine's states, inputs, outputs, etc. The view is going to show a diagram for the machine, graphically allowing the user to add new states, establish transitions, and put the states in a pleasing arrangement, among other things. I would like to store part of the diagram's state (e.g. the x and y state positions) when the machine information is stored for later retrieval, and am wondering how best to go about structuring the model(s?) for this. It seems like this UI information is more closely related to the view than to the state-machine model, so I was thinking that a secondary model might be in order, but I am reluctant to pursue this route because of the added complexity. Adding this information to the current model doesn't seem the right way to go about it either. This is the my first time using the MVC pattern so I'm still figuring things out. Any input would be appreciated.

    Read the article

  • Can I see if and when a file was deleted on Windows Server 2003?

    - by user316687
    On Windows Server 2003, is there a way to see if and when a file was deleted? It's a web server with IIS, our web application let our users to load Word documents into server. However, we found that one Word file is missing, and would like to know is it was deleted or never existed (web app could'nt load it). EDIT: I tried to follow this: Enable auditing the folder you want to keep track of. Just right click on the folder, go to “sharing and security”, then “security” tab, at the bottom click on “advanced”. Select the auditing tab, click add, select the group or users to track, then pick what actions you want to track. To track file deletion you would enable: Create files/Write data Success/Fail Create folders / append data Success/Fail Delete Subfolders/Files Success/Fail Delete Suceess/Fail This one will apply from now on, past actions wouldn't be able to track?

    Read the article

  • Include page url in php error_log

    - by scatteredbomb
    I'm working on sorting out all my php errors. I've just setup so that all errors go into one log so I can view them. The problem is that I can't really see what page someone trying to access that gave them an error. Of course I can see the file where the error was generated, but these are often includes used on other pages - so I can't really see what URL a person was attempting to view when this error was displayed, so I can't really replicate the problem. Any suggestions? Is it possible to include the actual page URL?

    Read the article

  • Accessing subfolders of a windows share from linux

    - by Born2Smile
    Hi, at my work they have a funny setup: my home folder is a subfolder to a share, as such: \\server\share\subfolder Now I have full permissions to the subfolder, but no permissions to share. From windows I can connect to the VPN of my work place, type the above address into any address field, and voila: I see the contents of my home folder. In Linux (using Ubuntu) however, I can't figure out how to connect directly to the subfolder. Every attempt I can think of keeps returning "Access denied", because I don't have permission to view the share. Any help on how to connect to the subfolder would be greatly appreciated :) Cheers, Born2Smile

    Read the article

  • How to mount a network drive?

    - by Relik
    Ok so I'm trying to set-up a home file server. I'm thinking about just setting it up as an FTP server, no particular reason other than I'm familiar with FTP and samba tends to be very frustrating. Basically the set-up I'm going for, is to be-able to create multiple user accounts for the server and restrict or allow access to specific folders on each user. FTP is the only way (that I know of) to accomplish a set-up like that. My question is how can I mount an FTP server as a drive in Ubuntu so that all my applications can access it just like any other driver or folder. An example would be downloading 12.10 via torrent when it comes out, I would like to be able to tell transmission to just download the file straight to my ftp server. I know how to do this in Windows, its actually very easy. But I cant figure it out in Ubuntu. I have tried using the "connect to server" option in nautilus, and it works, but it doesn’t give me the result I want, most applications don’t see the folder, while others can. Also I am open to options other than FTP if anyone has any suggestions. I've looked into FreeNAS but that doesn’t seem to allow me to control the user accounts the way I want to. Then after all is said and done I would still need a way to mount the shares as a drive in Ubuntu. The ability to mount network drives in windows is one of my favourite features, but seeing how Ubuntu is now my daily OS and has been for about 4 years, I really need a way to accomplish the same thing in Ubuntu. Also a GUI would be preferable, seeing as there will be multiple people using this server, I would like it to be as easy as possible. EDIT: this link here seems to be almost exactly what I'm wanting to do, if I could find a GUI that can do this ill be almost set. then I would just need to find a way to hide specific folders from certain users.

    Read the article

  • Best way to protect website application code

    - by Gaz_Edge
    Background I have a web application that I host on my own server. I have clients who use the application as is, but some have asked if they can host the application on their own server. This enables them to have their own URLS rather than mine. The application only forms part of their website so I'm assuming it will not be possible for my server to respond to a direct call to their domain etc To give some examples, i currently have urls like www.mydomain.com/profile, www.mydomain.com/index.php?option=someoption&view=someview&id=1 What my clients' want is www.theirdomian.com/profile, www.theirdomian.com/index.php?option=someoption&view=someview&id=1 etc Question My question is, what is the best way for me to allow them to use their own URLs with my application, without giving them all the backend source code and databases to install on their server? One way I thought would be to create a router.php file that sits on their server. The router then asks my server to output the html. The router modifies all the links etc in the html source and outputs the new html through the clients server. When a link is clicked on the clients site, the router receives the request and modifies the url to get the data from my server etc. Is this an effective way to achieve what I want, or is it way off the mark.

    Read the article

  • Architectural advice - web camera remote access

    - by Alan Hollis
    I'm looking for architectural advice. I have a client who I've built a website for which essentially allows users to view their web cameras remotely. The current flow of data is as follows: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Ftp connection is enabled for the cameras ftp user. Web camera opens ftp connection to server. Web camera begins taking photos. Web camera sends photo to ftp server. On image url request: Server reads latest image on hard drive uploaded via ftp for camera. Server deleted any older images from the server. This is working okay at the moment for a small amount of users/cameras ( about 10 users and around the same amount of cameras), but we're starting to worrying about the scalability of this approach. My original plan was instead of having the files read from the server, the web server would open up an ftp connection to the web server and read the latest images directly from there meaning we should have been able to scale horizontally fairly easily. But ftp connection establishment times were too slow ( mainly due to the fact that PHP out of the ox is unable to persist ftp connections ) and so we abandoned this approach and went straight for reading from the hard drive. The firmware provider for the cameras state they're able to build a http client which instead of using ftp to upload the image could post the image to a web server. This seems plausible enough to me, but I'm looking for some architectural advice. My current thought is a simple Nginx/PHP/Redis stack. Web camera issues post requests of latest image to Nginx/PHP and the latest image for that camera is stored in Redis. The clients can then pull the latest image from Redis which should be extremely quick as the images will always be stored in memory. The data flow would then become: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Camera is sent an http request to start posting images to a provided url Web camera begins taking photos. Web camera sends post requests to server as fast as it can On image url request: Server reads latest image from redis Server tells redis to delete later image My questions are: Are there any greater overheads of transferring images via HTTP instead of FTP? Is there a simple way to calculate how many potential cameras we could have streaming at once? Is there any way to prevent potentially DOS'ing our own servers due to web camera requests? Is Redis a good solution to this problem? Should I abandon PHP/Ngix combination and go for something else? Is this proposed solution actually any good? Will adding HTTPs to the mix cause posting the image to become too slow? Thanks in advance Alan

    Read the article

  • How to blacklist Terminal Services startup environment setting?

    - by JBurace
    I have a user in Active Directory who uses this setting in the Environment tab: Start the following program at logon: "C:\Program Files\PName\Folder\gui.exe" This runs okay on various computers (that are on the domain) including his own. But the user needs to RDP into a Windows Server which does not have this program (which is normal). When the user RDPs into the server and logs in with the AD account, an error occurs about C:\Program Files\PName\Folder\gui.exe missing and the user then gets stuck at a grey screen. The user needs to RDP into this server; how can one blacklist that Environment setting from activation on a specific machine on the domain?

    Read the article

  • Which permissions need to be assigned to a normal user to allow him to change SHARE permissions?

    - by guest
    As in the subject - I am wondering, which permissions (if it is possible) do I need to assign to a regular user to modify the share permissions - e.g. to add another user with full control permissions, or to deny someone read access to the folder - on a share level. I know that this is possible through NTFS permissions, but I am wondering whether it is also possible on share level. Any ideas how to do that? Or perhaps only the adminstrator/creator/person who shared the folder has the access to this? I am using win2003 Any ideas? Thanks

    Read the article

  • Keep IMAP messages locally when deleted remotely

    - by user74328
    I use my email from my phone and my computer via IMAP. I want to set something up so that if I delete a message via my phone, my computer will still keep the message locally. For example, assume I leave my computer on, with a synchronize interval of 5 minutes. I want to be able to send something from my phone, wait 5 minutes to be sure my computer has downloaded the item from the Sent folder, then delete it from the IMAP sent folder via my phone, but have the computer at home keep it's copy. Is this possible with any readily available email clients out there? I have Thunderbird and Outlook at the moment, but would be willing to learn a new interface for this feature. How can I accomplish this?

    Read the article

< Previous Page | 485 486 487 488 489 490 491 492 493 494 495 496  | Next Page >