Search Results

Search found 24734 results on 990 pages for 'floating point conversion'.

Page 541/990 | < Previous Page | 537 538 539 540 541 542 543 544 545 546 547 548  | Next Page >

  • Can not find the source of Grant permission on a folder

    - by Konrads
    I have a security mystery :) Effective permissions tab shows that a few sampled users (IT ops) have any and all rights (all boxes are ticked). The permissions show that Local Administrators group has full access and some business users have too of which the sampled users are not members of. Local Administrators group has some AD IT Ops related groups of which the sampled users, again, appear not be members. The sampled users are not members of Domain Administrators either. I've tried tracing backwards (from permissions to user) and forwards (user to permission) and could not find anything. At this point, there are two options: I've missed something and they are members of some groups. There's another way of getting full permissions. Effective Permissions are horribly wrong. Is there a way to retrieve the decision logic of Effective Permissions? Any hints, tips, ideas?

    Read the article

  • How to access MySQL on Windows

    - by Dan
    This may sound like a really dumb question, but I normally only deal with MSSQL, no LAMP stuff, so I'm struggling to figure out what's going on. I have Windows 7 and have installed MySQL 5.1 through Web Platform Installer. I have HeidiSQL installed to manage data in MySQL, but how do I connect? In Heidi it's asking for 'Hostname / IP' which is prepopulated with 127.0.0.1. It prepopulates the user field to 'root' (which is right) and I'm entering the password I chose when MySQL was installed. However, it just errors when I connect, saying: SQL Error (1045): Access denied for user 'root'@'localhost' (using password: YES). Can anyone point me in the right direction here? Many thanks...

    Read the article

  • Multiple Issues - USB booting Ubuntu 12.04

    - by Pixelishus
    I've been using a Ubuntu 12.04 off a bootable USB stick so I can have a portable OS. Also, OS variety. I've been using it for a while and I'm still have a few issues. First and most annoying, it often lags/freezes. I'm not sure what word you would use for it, windows will often stop working and go dark/not responding from anywhere between 10 seconds to a full minute and on some occasions, even longer. However, the window eventually starts working again. Similarly, sometimes the whole system will freeze, not just the window. The mouse will still move, but nothing will work. No clicking, no menus, not keyboard shortcuts. Again, it will usually start working. I'm liking Ubuntu a lot, but those issues can make it annoying to use sometimes. For example, it will ALWAYS freeze at some point if I try to watching a Youtube video and I'll have to wait for a minute or so until it starts responding, again. Aside from the lag/freezing, anytime I download packages, it will always say "package operation failed" when it's done, though it does seem to download/install. Another issue I'm having is with shutting down. If I open the logout/shutdown menu and click shutdown, it just logs me out and takes me to the login screen. If I try shutting down through the login screen, it won't do anything. As if I didn't even go to it. I've been using the terminal to reboot or shutdown when I need to. I've looked around for answers for all of these problems and still have yet to find a solution that works. Are these just normal issues with USB booting? I haven't installed Ubuntu to any computers, I've always done USB booting.

    Read the article

  • How to define implementation details?

    - by woni
    In our project, an assembly combines logic for the IoC-Container, the project internals and the communication layer. The current version evolved to have only internal classes in addin assemblies. My main problem with this approach is, that the entry point is only available over the IoC-Container. It is not possible to use anything else than reflection to initialize the assembly. Everything behind the IoC-Interface is defined as implementation detail and therefore not intended for usages outside. It is well known that you should not test implementation detail (such as private and internal methods), because they should be tested through the public interface. It is also well known, that your tests should not use the IoC-Container to setup the SUTs, because that would result in too much dependencies. So we are using the InternalsVisibleTo-Attribute to make internals visible to our test assemblies and test the so called implementation details. I recognized that one problem could be the mixup between different concerns in that assembly, changing this would make this discussion useless, because classes have to be defined public. Ignoring my concerns with this, isn't the need to test a class enough reason to make it public, the usages of InternalsVisibleTo seems unintended, and a little bit "hacky". The approach to test only against the publicly available IoC-Container is too costly and would result in integration style tests. The pros of using internals are, that the usages are well known and do not have to be implemented like a public method would have to be (documentation, completeness, versioning,...). Is there a solution, to not test against internals, but keep their advantages over public classes, or do we have to redefine what an implementation detail is.

    Read the article

  • Lost Root and other user passwords

    - by Webnet
    This isn't a huge deal, because there's very little on the server (literally a file or two) that we actually need off of it. But we disabled root logins as a security measure and can't remember any of our other user passwords. I'm assuming that there's nothing we can do at this point to get into the server? I'm sitting next to the box... Update Oops... actually, I need to export an SVN off of this server. So yeah, there's stuff I need.

    Read the article

  • MariaDB, Galera, xtrabackup - do I need the binary log?

    - by bernhardrusch
    We are using a MariaDB Galera Cluster with 3 nodes. For the state transfer we are using xtrabackup. We have some problems with the binary logs - they got too big and crashed the server. We can remove them manually with the purge binary logs command, another way would be to set the expire_logs_days so they would expire. I now that we could use xtrabackup to backup the DB and use the binlog to get to some point in time. But do we really need it for Galera to work ?

    Read the article

  • Should I expect my peers to read or practice on a regular basis? [closed]

    - by Joshua Smith
    I've been debating asking this question for some time. Based several of the comments I read in this question I decided I had to ask. This feels like I'm stating the obvious, but I believe that regular reading (of books, blogs, StackOverflow, whatever) and/or practice are required just to stay current (let alone excel) in whichever stack you use to pay the bills, not to mention playing with things outside your comfort zone to learn new ways of doing things. Yet, I virtually never see this from many of my peers. Even when I go out of my way to point out useful (and almost always free) learning material, I quite often get a sense of total apathy from those I'm speaking to. I'd even go so far as to say that if someone doesn't try to improve (or at least stay current), they'll atrophy as technology advances and actually become less useful to the company. I don't expect people to spend hours a day studying or practicing. I have two young kids and hours of practice simply aren't feasible. Still, I find some time; perhaps on the train, at lunch, in bed for a few minutes, whatever. I'm willing to believe this is arrogance or naivete on my part, but I'd like to hear what the community has to say. So here's my question: Should I expect (and encourage) the same from my peers, or just keep my mouth shut and do my own thing?

    Read the article

  • Why isn't one of the constant buffers being loaded inside the shader?

    - by Paul Ske
    I however got the model to load under tessellation; only problem is that one of the constant buffers aren't actually updating the shader's tessellation factor inside the hullshader. I created a messagebox at the rendering point so I know for sure the tessellation factor is assigned to the dynamic constant buffer. Inside the shader code where it says .Edges[1] = tessellationAmount; the tessellationAmount is suppose to be sent from the dynamic buffer to the shader. Otherwise it's just a plain box. In better explanation; there's a matrixBuffer, cameraBuffer, TessellationBuffer for constant. There's a multiBuffer array that assigns the matrix, camera, tesselation. So, when I set the Hull Shader, PixelShader, VertexShader, DomainShader it gets assigned by the multibuffer. E.G. devcon-HSSetConstantBuffers(0,3,multibuffer); The only way around the whole ideal would be to go in the shader and change how much the edges tessellate and inside the edges as well with the same number. My question is why wouldn't the tessellationBuffer not work in the shader?

    Read the article

  • Linksys Wireless-N ADSL2 WAG160N losing connection

    - by Sam1973
    It seems that recently I have to turn off and then turn on my router every time I want to connect my laptop(s) to the internet. This is what's happens: When I want to connect to the internet on my laptop, the network icon in the system tray states that I am connected to the wireless network but there's no "internet connectivity" I turn off the router and turn it back on and then I can connect to the internet. I am sure there is something that needs to be done. Can you guys point me to the right direction?

    Read the article

  • Installer hangs on "waiting for other installations to finish" despite no other installations in progress

    - by clozach
    I started an installation of the Xcode 4 developer's preview last night before going to bed. When I woke up, the installation had popped up a message saying to close Xcode first (doh!). I did so, but whatever script had displayed the message seemed to have hung because the (uncancellable) message wouldn't disappear, and the installer wouldn't progress. After force quitting the Installer app, the message above was still showing, so I restarted my MacBook Pro and then tried the installation again. This time it wouldn't even start, instead showing one of those candy-stripe activity bars and the message: Waiting for other installations to finish At this point, I'm stumped. There are no other installations in progress, so there must be some artifact left over from the interrupted install that's gumming up the works. But what?

    Read the article

  • Capture the build number for a remote-triggered Hudson job?

    - by EMiller
    I have a very simple inhouse web app from which certain Hudson builds (on another server) can be triggered remotely. I have no problem triggering the builds, but I don't know how to capture the associated build number for later reference. I'm using the buildWithParameters trigger, and the actual result of that call is just a mess of HTML - I don't believe it gives me back the build number. I started down the path of pulling the whole build list for the job (via the api), and then attempting to reconcile that list against my records - but that's much more complicated than I'd like it to be. I also considered sleeping for a few seconds after launching the job, and then grabbing the latestBuild from the Hudson api - but I'm sure that's going to go wrong at some point (someone will fire off two jobs quickly, and I'll get the association wrong).

    Read the article

  • Fresh install of 64 bit 12.04 over 32 bit 11.10 alongside Windows 7

    - by Pareen
    I currently have Ubuntu 11.10 32 bit and Windows 7 dual boot in separate partitions. I am trying to do a fresh install of Ubuntu 12.04 64 bit (mistakenly installed the 32 bit 11.10 a little while ago.. I need a 64 bit version to support AOSP build) OVER my the exisiting 11.10 partition. I have referenced How to Install fresh 12.04 install to a PC with dual booting Windows 7 & Ubuntu11.10?, as well as other posts on using the Live CD to do a fresh install. However, the problem I am experiencing is when I bring up the install screen, it says the following: This computer has multiple operating systems on it. What would you like to do. (3 options) Install Ubuntu 12.04 alongside them Replace all with Ubuntu 12.04 (Warning, this will delete files across ALL operating systems) Something else (you can create or resize partitions yourself) This is different from what is in other posts, as mine states that there are "multiple O.Ses" and doesnt individually allow me to replace the Ubuntu 11.10. I don't want to replace ALL O.S.es: I need to preserve Windows 7 and am only trying to replace the old Ubuntu 11.10 partition with the new 12.04 64 bit. I did have Ubuntu installed via Wubi (I believe it was 10.04) prior to putting 11.10 in a separate partition, but I have removed it via Add/Remove programs in Windows. I was wondering how to go about doing this... Should I use the "Something else" option to bring up the partition manager, and just assign my existing 11.10 partition with root mount point + swap space. Will this do the same thing by overwriting with fresh 12.04 install?? I appreciate all your help.

    Read the article

  • Win7 XP mode install fails with cryptic error

    - by MAW74656
    Trying to install XPmode on Win7 Professional 64-bit SP1, but getting an error. I downloaded XPMode, Virtual PC and the update as specified here: http://www.microsoft.com/windows/virtual-pc/download.aspx. When I open XP mode for the first time it goes into the setup ("Setting up Windows XP Mode for first use..."; pick your VM password, etc.), then it runs through the progress bar until almost the end (99%?). At that point I get the error: Could not complete Setup. Please try again. Whats going on here? I've uninstalled Virtual PC and xpmode, and reinstalled in several different orders, no luck. Is there some log file that will tell me whats happening? Theres nothing showing up in the Event log that I can find. How can I fix this?

    Read the article

  • How to avoid jumping to a solution when under pressure? [closed]

    - by GlenPeterson
    When under a particularly strict programming deadline (like an hour), if I panic at all, my tendency is to jump into coding without a real plan and hope I figure it out as I go along. Given enough time, this can work, but in an interview it's been pretty unsuccessful, if not downright counter-productive. I'm not always comfortable sitting there thinking while the clock ticks away. Is there a checklist or are there techniques to recognize when you understand the problem well enough to start coding? Maybe don't touch the keyboard for the first 5-10 minutes of the problem? At what point do you give up and code a brute-force solution with the hope of reasoning out a better solution later? When is it most productive to think and design more vs. code some experiments to and figure out the design later? Here is a list of techniques for taking a math test and another for taking an oral exam. Is there is a similar list of techniques for handling a programming problem under pressure? ANSWERS: I think this is a valid answer: How To Solve It. I found the link as an answer to Steps to solve or approach towards a solution. There were also some really good tips at Is thinking out loud during an interview really the best strategy?. A great and concise argument for TDD is the first answer to TDD Writing code vs Figuring out the answer to a problem?. My question may be a near-duplicate of that one.

    Read the article

  • How can HAProxy improve availibility, or "how can I prevent my site from going down"? [closed]

    - by Joe Hopfgartner
    I am aware of what HAProxy does, but what if my HAproxy goes down? Or what if my DNS servers go down? Yes, dns is less the problem. However dns only solves to an IP and an IP is announced via BGP to be routed over some router. What if that router goes down? Of course if I have complicated application servers that are likely to fail HAProxy can significantly improve uptime. But my application isnt. In fact my application may very well just be delivering a small static html file via HTTP. Basically if any user anywhere types in MYDOMAIN.COM, I want the user to get SOMETHING on the screen other than a timeout or DNS resolution error. How can I do that? The point of entry is difficult. The so called "initial closure mechanism".

    Read the article

  • Set a 2d camera's position so that a building is under the players feet

    - by Potato
    My issue is this: i am making a scrolling game in XNA, and the camera updates based on the players velocity, but the player never actually moves, he is always in the center of the screen. When he hits the top of the building though i want him to always be on top and sink through the texture in a way like this: what i am doing to make this happen is i am just setting his velocity to 0, so its not moving, but the more velocity he hits a building with the more he sinks through it. I also tried setting the buildings position to the plays Bounding Box's bottom, and this achieved the look i wanted but this also resulted in the other buildings rising in the air, because the velocity was still moving (even if i set it to 0). if it was not a scrolling game, this would be not a problem, because you just set the players position to the top of the building, but because the player never actually moves, i actually need to move the camera to the point where the building is under the players feet without the other buildings rising. (Take note this is note a real camera, it is just a class that moves the objects in the world based on the players velocity). All questions are welcome.

    Read the article

  • How to host ASP.NET application externally?

    - by Josh
    I have an ASP.NET application that I can get to locally by going to 192.168.1.102:81/TestApp. I would like to host the application externally by going to domain.com:81/TestApp (I already have my domain pointing to my router and this works fine - I have apache running on port 80 on another server). I modified the router settings to point any request coming through port 81 to 192.168.1.102. I am still having trouble accessing the ASP.NET site (I get the error message that "This link appears to be broken"). Am I missing something? How can I redirect domain.com:81/TestApp to my ASP.NET application? Thanks.

    Read the article

  • ubuntu 12.04 installer does not recognize drive partitions

    - by endless forms
    I recently purchased a new HP Pavilion HPE desktop running Windows 7. I am trying to install a dual-boot system with 12.04. However, when I run the LiveCD I only get as far as the "Install" window where you can select the partitions for your drives. On the bottom where it says "device for boot loader installation" I have "/dev/sda" and cannot select any other devices. All the options to change the drives are greyed out, most likely because there are no drives in the window. I partitioned my largest drive using the tools within Windows, then booted into the CD, but nothing shows up. I then used Gparted to change the new space from unallocated to an /ext2, and still nothing shows up. The installer does not recognize anything, but when I go into an Ubuntu session and use the disk utility manager I can see the partitions I made. Anything I do has to be done outside of the installer. I have no files on this new computer, so this is the perfect time to install a parallel OS. I would like avoid completely reinstalling Windows, however. I've been over the forums many times, but all the answers I've found have not worked for me. I also tried flagging the new, empty partition as boot, but that screwed Windows up. Also, the WUBI installer hits the same point and quits. I know that the disk itself is fine because I just made another dual boot system on a Gateway PC. This makes me think something within this computer is preventing the installer from "seeing" the drives. Any help would be much appreciated! Edit in response: The main part of the partitioning window shows no partitions, everything is blank. There is no way to add partitions, and all the buttons are useless. I've tried defragging my drive multiple times, and I also used the same disk to dual-boot another PC with no problems, so it's not the disk, it's definitely the computer.

    Read the article

  • Automate Windows 8.1 Enterprise upgrade [migrated]

    - by Ben M.
    I have been trying to find the necessary command line switches for automating the upgrade for my Windows 8 clients to Windows 8.1. I have the ISO extracted and I've run setup.exe /? but that doesn't tell me enough. I can't find any relevant information from search engines. Can anyone point me to some documentation or information on how to automate the upgrade so that it keeps user data, programs, etc? I know how to do it when running the installer manually, but I obviously do not wish to do that with 100+ machines.

    Read the article

  • Problem with domain getting turned to IP address for https

    - by user229133
    I have a website that is using Windows Server 2003. The site is called https://mysite.com/ and at ip address 111.1.1.1. Now when I log into the site all my relative links that are generated using NavURL (<%# NavURL("Images/Menu/img.gif")%) are saying "http://111.1.1.1/Images/Menu/img.gif" instead of "https://mysite.com/Images/Menu/img.gif". This is causing an error because it needs to be secure. I'm sure there is a setting on the server somewhere to point to the name and not the ip, but I don't know where. Thanks for your help.

    Read the article

  • htaccess and htpasswd trouble

    - by hjpotter92
    This is the first time that I have ever tried working with .htpasswd and .htaccess files, so please point out my childish works. I have my apache document root set to /www/ on my debian server. Inside it, there's a folder named Logs/ which I want to restrict access using a htpasswd. I created my htpasswd file using the shell's htpasswd command. And this is the result: > user:<encoded password here> > hjp:<encoded password here> > hjpotter92:<encoded password here> I put this file named .htaccess inside /www/. The Logs/ has following htaccess file in it: > AuthName "Restricted Area" > AuthType Basic > AuthUserFile /www/.htpasswd > AuthGroupFile /dev/null > require valid-user This was again created using an online tool(I forgot its name/link, and can't search the browser-history now). The problem, as it might've already struck you is that I am experiencing no change on my Logs folder access. The folder is still accessible to everyone. I am running apache as root user(if that matters/helps). Please help/guide me. I've tried reading some htaccess guides and have followed some of older SO questions, but still haven't figured out a way to restrict access to Logs folder with a password.

    Read the article

  • Browsing is much slower on one PC wired to the same router - why?

    - by deanalt
    Wife is not happy. It takes about 5 seconds to open a google window, versus about 1 second on the faster computer which is about 3 years old itself. Yes, it is an older computer (5 -6 years old, I'd guess), surely with less RAM, but for simple browsing, should it matter? Both are hardwired to the same Netgear Rangemax router. Both use fixed IP addresses. Both are XP. Both have about 8 feet of cable to the router. I have the fastest service my cable provides. Probably irrelevant but ...two newer MACs are connected wirelessely during the summer and they are even faster, but I think that's the difference in browsers. If you could point me to a list of process of elimination steps that would be most appreciated. Thanks Dean

    Read the article

  • Who should have full visibility of all (non-data) requirements information?

    - by ebyrob
    I work at a smallish mid-size company where requirements are sometimes nothing more than an email or brief meeting with a subject matter manager requiring some new feature. Should a programmer working on a feature reasonably expect to have access to such "request emails" and other requirements information? Is it more appropriate for a "program manager" (PGM) to rewrite all requirements before sharing with programmers? The company is not technology-centric and has between 50 and 250 employees. (fewer than 10 programmers in sum) Our project management "software" consists of a "TODO.txt" checked into source control in "/doc/". Note: This is nothing to do with "sensitive data access". Unless a particular subject matter manager's style of email correspondence is top secret. Given the suggested duplicate, perhaps this could be a turf war, as the PGM would like to specify HOW. Whereas WHY is absent and WHAT is muddled by the time it gets through to the programmer(s)... Basically. Should specification be transparent to programmers? Perhaps a history of requirements might exist. Shouldn't a programmer be able to see that history of reqs if/when they can tell something is hinky in the spec? This isn't a question about organizing requirements. It is a question about WHO should have full VISIBILITY of requirements. I'd propose it should be ALL STAKEHOLDERS. Please point out where I'm wrong here.

    Read the article

  • Windows 7 with two network cards doesn't route traffic

    - by Tomek
    I have simple task to do: I have wni7 with two nics.I want to connect another comp(osx) to win7 through second nic to connect it to internet. I already changed the registry. Win7 interface with 192.168.2.1 has no gateway set (no point to do that) OSX interface with 192.168.2.2 has gateway set to 192.168.2.1 I do not add any routes on win7, every thing seems to be already there network on second nic is detected as "undefined network" (probably effect of no gateway) i can achieve any connectivity to internet from OSX only by enabling network connection sharing on nic with 192.168.2.1, but it enables NAT and I'm interested only in pure routing without nat(it's a setup for some research). firewall is off. It seems to me that win7 refuses to forward packets for some reason. Perhaps "undefined network" and NLA service is to blame, although i couldn't find any info about that. Below ascii schematics of my setup: internet<--router(192.168.1.1)<--(192.168.1.1) WIN7 (192.168.2.1)<--(192.168.2.2)OSX Thanks

    Read the article

  • mod_proxy failing as forward proxy in simple configuration

    - by Stabledog
    (On Mac OS X 10.6, Apache 2.2.11) Following the oft-repeated googled advice, I've set up mod_proxy on my Mac to act as a forward proxy for http requests. My httpd.conf contains this: <IfModule mod_proxy> ProxyRequests On ProxyVia On <Proxy *> Allow from all </Proxy> (Yes, I realize that's not ideal, but I'm behind a firewall trying to figure out why the thing doesn't work at all) So, when I point my browser's proxy settings to the local server (ip_address:80), here's what happens: I browse to http://www.cnn.com I see via sniffer that this is sent to Apache on the Mac Apache responds with its default home page ("It works!" is all this page says) So... Apache is not doing as expected -- it is not forwarding my browser's request out onto the Internet to cnn. Nothing in the logfile indicates an error or problem, and Apache returns a 200 header to the browser. Clearly there is some very basic configuration step I'm not understanding... but what?

    Read the article

< Previous Page | 537 538 539 540 541 542 543 544 545 546 547 548  | Next Page >